zecanard's picture
Add files using upload-large-folder tool
5410f6a verified
---
base_model: trohrbaugh/gemma-4-31b-it-heretic-ara
language: en
library_name: mlx
license: apache-2.0
license_link: https://ai.google.dev/gemma/docs/gemma_4_license
pipeline_tag: image-text-to-text
tags:
- mlx
- heretic
- uncensored
- decensored
- abliterated
- ara
---
# 🦆 zecanard/gemma-4-31b-it-uncensored-heretic-ara-MLX-5bit-int5-affine
[This model](https://huggingface.co/zecanard/gemma-4-31b-it-uncensored-heretic-ara-MLX-5bit-int5-affine) was converted to MLX from [`trohrbaugh/gemma-4-31b-it-heretic-ara`](https://huggingface.co/trohrbaugh/gemma-4-31b-it-heretic-ara) using `mlx-vlm` version **0.5.0**.
Please refer to the [original model card](https://huggingface.co/trohrbaugh/gemma-4-31b-it-heretic-ara) for more details.
## 🌟 Quality
Quantized vision language model with an effective **7.852 bits per weight**.
`mlx_vlm.convert --quantize --q-group-size 32 --q-bits 5 --q-mode affine`
## 🛠️ Customizations
This quant is aware of the current date, and also enables thinking (if available). You may disable this behavior by deleting the following line from the chat template, or changing `true` to `false`:
`{%- set enable_thinking = true %}`
You may also need to adjust your environment’s **Reasoning Section Parsing** to recognize `<|channel>thought` as the **Start String**, and `<channel|>` as the **End String**.
## 🖥️ Use with `mlx`
```bash
pip install -U mlx-vlm
```
```bash
mlx_vlm.generate --model zecanard/gemma-4-31b-it-uncensored-heretic-ara-MLX-5bit-int5-affine --max-tokens 100 --temperature 0 --prompt "Describe this image." --image <path_to_image>
```