Quantized version of darkc0de/Hermes-4.3-36B-heretic. This is an abliteration method that doesn't seem to be as damaging to the base model as the previous methods were.
The repo includes the following quantized file:
For cards with 24GB of VRAM
- IQ4_XS (with iMatrix): This should allow you to run this with 24GB VRAM at 16K context length.
Settings
Instruction Template: Llama3-Chat Thinker
See Hermes's Official Readme for complete documentation and references.
- Downloads last month
- 29
Hardware compatibility
Log In to add your hardware
4-bit
Model tree for SerialKicked/Hermes-4.3-36B-heretic-GGUF-IQ4_XS
Base model
darkc0de/Hermes-4.3-36B-heretic