MLX-Qwopus3.5-27B-v3-vision-6bit

A 6-bit MLX quantization of Jackrong/Qwopus3.5-27B-v3 with a few tweaks to restore the multimodal capabilities.

Supports {%- set enable_thinking = false %} Jinja variable.

Update [2026-04-12]: Refined the chat template to further improve stability for long-running tasks and tool use; mitigated an issue where incorrect think tag formatting could leak from a distillation dataset.


Quantization Details

Property Value
Method 6-bit (6.661 bits per weight)
Tool mlx-vlm 0.4.2 via mlx-vlm.convert
Size ~22.9GB

Other Available Quants

Model Size Quantization Bits per weight Multimodal
Jackrong/MLX-Qwopus3.5-27B-v3-4bit 15.15 GB 4-bit 4.501
matt-here/MLX-Qwopus3.5-27B-v3-vision-4bit 16.08 GB 4-bit 4.695 ✓ (Vision)
matt-here/MLX-Qwopus3.5-27B-v3-5bit 18.56 GB 5-bit 5.501
matt-here/MLX-Qwopus3.5-27B-v3-vision-5bit 19.46 GB 5-bit 5.678 ✓ (Vision)
Jackrong/MLX-Qwopus3.5-27B-v3-6bit 21.88 GB 6-bit 6.501
(This model) 22.85 GB 6-bit 6.661 ✓ (Vision)
Jackrong/MLX-Qwopus3.5-27B-v3-bf16 53.81 GB bf16 16

GGUF quants - Jackrong/Qwopus3.5-27B-v3-GGUF


Credits

Downloads last month
649
Safetensors
Model size
6B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for matt-here/MLX-Qwopus3.5-27B-v3-vision-6bit

Base model

Qwen/Qwen3.5-27B
Adapter
(11)
this model