• Original model: aifeifei798/Gemma-4-31B-Cognitive-Unshackled - refer for more details on the model.
  • This is a backup quant inferior to mradermacher/Gemma-4-31B-Cognitive-Unshackled-i1-GGUF. I recommend: Gemma-4-31B-Cognitive-Unshackled.i1-IQ4_XS.gguf as a replacement.

Okay. So I ran this in llama.cpp and Silly Tavern chat completion. On my 24VRAM this fits 32000 context at f16 kv cache and 1024 batch. Works fine, long RP context coherence, thinking and no thinking. No breaking, no issues.

Example Dialogue

Example dialogue

Downloads last month
1,685
GGUF
Model size
31B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for s1arsky/Gemma-4-31B-Cognitive-Unshackled-Q4_KS_GGUF

Quantized
(3)
this model