matt-here's picture
Update README.md
22a2415 verified
|
raw
history blame
2.49 kB
metadata
language:
  - en
  - zh
  - ko
license: apache-2.0
base_model: Jackrong/Qwopus3.5-27B-v3
tags:
  - unsloth
  - qwen
  - qwen3.5
  - reasoning
  - chain-of-thought
  - lora
  - competitive-programming
  - mlx
pipeline_tag: image-text-to-text
library_name: mlx

MLX-Qwopus3.5-27B-v3-vision-4bit

A 4-bit MLX quantization of Jackrong/Qwopus3.5-27B-v3 with a few tweaks to restore the multimodal capabilities.

Supports {%- set enable_thinking = false %} Jinja variable.

Update [2026-04-12]: Refined the chat template to further improve stability for long-running tasks and tool use; mitigated an issue where incorrect think tag formatting could leak from a distillation dataset.


Quantization Details

Property Value
Method 4-bit (4.695 bits per weight)
Tool mlx-vlm 0.4.2 via mlx-vlm.convert
Size ~16.1GB

Other Available Quants

Model Size Quantization Bits per weight Multimodal
Jackrong/MLX-Qwopus3.5-27B-v3-4bit 15.15 GB 4-bit 4.501 ✗
(This model) 16.08 GB 4-bit 4.695 ✓ (Vision)
matt-here/MLX-Qwopus3.5-27B-v3-5bit 18.56 GB 5-bit 5.501 ✗
matt-here/MLX-Qwopus3.5-27B-v3-vision-5bit 19.46 GB 5-bit 5.678 ✓ (Vision)
Jackrong/MLX-Qwopus3.5-27B-v3-6bit 21.88 GB 6-bit 6.501 ✗
matt-here/MLX-Qwopus3.5-27B-v3-vision-6bit 22.85 GB 6-bit 6.661 ✓ (Vision)
Jackrong/MLX-Qwopus3.5-27B-v3-bf16 53.81 GB bf16 16 ✗

GGUF quants - Jackrong/Qwopus3.5-27B-v3-GGUF


Credits