- Original model:
aifeifei798/Gemma-4-31B-Cognitive-Unshackled- refer for more details on the model. - This is a backup quant inferior to mradermacher/Gemma-4-31B-Cognitive-Unshackled-i1-GGUF. I recommend: Gemma-4-31B-Cognitive-Unshackled.i1-IQ4_XS.gguf as a replacement.
Okay. So I ran this in llama.cpp and Silly Tavern chat completion. On my 24VRAM this fits 32000 context at f16 kv cache and 1024 batch. Works fine, long RP context coherence, thinking and no thinking. No breaking, no issues.
Example Dialogue
- Downloads last month
- 1,685
Hardware compatibility
Log In to add your hardware
4-bit
Model tree for s1arsky/Gemma-4-31B-Cognitive-Unshackled-Q4_KS_GGUF
Base model
google/gemma-4-31B-it