Mar 5 Update: New iMatrix + Variants

#20
by danielhanchen - opened
Unsloth AI org
β€’
edited Mar 5
  • All GGUFs now use our new imatrix data. See some improvements in chat, coding, long context, and tool-calling use-cases.
  • Rest of variants like UD_K_XL are now uploaded.
  • Replaced BF16 layers with F16 for faster inference on unsupported devices.
  • Think toggle for Qwen3.5 now in LM Studio. See our guide for instructions.
  • Please follow the correct instructions / settings in our guide here.

Note: This may not be our final update to 27B.

Fine-tuning and RL Qwen3.5

danielhanchen pinned discussion

Just to confirm: 27B is still updating or already updated? I mean should we wait until we see at least an upload date of 5th of March or later, or was it already updated on 2nd of March?

Note: This may not be our final update to 27B.

-) Have you seen that Qwen released a 4bit quant two days ago? https://huggingface.co/Qwen/Qwen3.5-27B-GPTQ-Int4/blob/main/config.json#L135
-) Would there be any benefit to use a domain-optimized dataset for imatrix quants to preserve higher model fidelity in that domain? e.g., agentic coding, tool calling, reasoning, STEM. or is this more a holistic process where you cannot neglect or prefer?

Just to confirm: 27B is still updating or already updated? I mean should we wait until we see at least an upload date of 5th of March or later, or was it already updated on 2nd of March?

from a comment by daniel on r/localllama
"27B for now works reasonably ok - but we do plan to update them over the weekend - we're still getting bottlenecked by 397B :( "

thanks for surfacing the updated info for us.

So is the 27B still getting updated? Files are still the old ones... At least remove the March 5th update on the page if you don't plan to do this anymore...

Sign up or log in to comment