Upload 3 files
Browse files- .gitattributes +2 -0
- Gemma-4-Garnet-31B-it-mmproj-BF16.gguf +3 -0
- Gemma-4-Garnet-31B-it-uncensored-heretic-BF16.gguf +3 -0
- README.md +194 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
Gemma-4-Garnet-31B-it-mmproj-BF16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
Gemma-4-Garnet-31B-it-uncensored-heretic-BF16.gguf filter=lfs diff=lfs merge=lfs -text
|
Gemma-4-Garnet-31B-it-mmproj-BF16.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a3c3021444a0384b630396c9e2cd0c4b7ab9f9358d9612f66b7dbf5372142519
|
| 3 |
+
size 1200726080
|
Gemma-4-Garnet-31B-it-uncensored-heretic-BF16.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2b6110c4bbd7c7b7c63d4a9814d3f3f4362e27fba23f7a997b14c8d574a54f62
|
| 3 |
+
size 61413187328
|
README.md
ADDED
|
@@ -0,0 +1,194 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
base_model:
|
| 4 |
+
- ConicCat/Gemma4-Garnet-31B
|
| 5 |
+
pipeline_tag: text-generation
|
| 6 |
+
datasets:
|
| 7 |
+
- ConicCat/Gutenberg-SFT
|
| 8 |
+
- ConicCat/Condor-SFT-Filtered
|
| 9 |
+
tags:
|
| 10 |
+
- heretic
|
| 11 |
+
- uncensored
|
| 12 |
+
- decensored
|
| 13 |
+
- abliterated
|
| 14 |
+
- ara
|
| 15 |
+
---
|
| 16 |
+
<div style="background-color: #ff4444; color: white; padding: 20px; border-radius: 10px; text-align: center; margin: 20px 0;">
|
| 17 |
+
<h2 style="color: white; margin: 0 0 10px 0;">🚨⚠️ I HAVE REACHED HUGGING FACE'S FREE STORAGE LIMIT ⚠️🚨</h2>
|
| 18 |
+
<p style="font-size: 18px; margin: 0 0 15px 0;">I can no longer upload new models unless I can cover the cost of additional storage.<br>I host <b>70+ free models</b> as an independent contributor and this work is unpaid.<br><b>Without your support, no more new models can be uploaded.</b></p>
|
| 19 |
+
<p style="font-size: 20px; margin: 0;">
|
| 20 |
+
<a href="https://patreon.com/LLMfan46" style="color: white; text-decoration: underline;">🎉 Patreon (Monthly)</a> |
|
| 21 |
+
<a href="https://ko-fi.com/llmfan46" style="color: white; text-decoration: underline;">☕ Ko-fi (One-time)</a>
|
| 22 |
+
</p>
|
| 23 |
+
<p style="font-size: 16px; margin: 10px 0 0 0;">Every contribution goes directly toward Hugging Face storage fees to keep models free for everyone.</p>
|
| 24 |
+
</div>
|
| 25 |
+
|
| 26 |
+
---
|
| 27 |
+
|
| 28 |
+
### **94% fewer refusals** (6/100 Uncensored vs 99/100 Original) while preserving model quality (0.0368 KL divergence).
|
| 29 |
+
|
| 30 |
+
## ❤️ Support My Work
|
| 31 |
+
Creating these models takes significant time, work and compute. If you find them useful consider supporting me:
|
| 32 |
+
|
| 33 |
+

|
| 34 |
+
|
| 35 |
+
| Platform | Link | What you get |
|
| 36 |
+
|----------|------|--------------|
|
| 37 |
+
| 🎉 Patreon | [Monthly support](https://patreon.com/LLMfan46) | Priority model requests |
|
| 38 |
+
| ☕ Ko-fi | [One-time tip](https://ko-fi.com/llmfan46) | My eternal gratitude |
|
| 39 |
+
|
| 40 |
+
Your help will motivate me and would go into further improving my workflow and coverings fees for storage, compute and may even help uncensoring bigger model with rental Cloud GPUs.
|
| 41 |
+
|
| 42 |
+
-----
|
| 43 |
+
|
| 44 |
+
GGUF quantizations of [llmfan46/Gemma-4-Garnet-31B-it-uncensored-heretic](https://huggingface.co/llmfan46/Gemma-4-Garnet-31B-it-uncensored-heretic).
|
| 45 |
+
|
| 46 |
+
This model is great for creative writing and translations, the original base model writing and translations feels very stiff with some odd word choices that might not really fit very well the situation, Gemma-4-Garnet-31B-it-uncensored-heretic aims to fix this issue and improve the writing quality of Gemma 4 31B it.
|
| 47 |
+
|
| 48 |
+
# This is a decensored version of [ConicCat/Gemma4-Garnet-31B](https://huggingface.co/ConicCat/Gemma4-Garnet-31B), made using [Heretic](https://github.com/p-e-w/heretic) v1.2.0 with the [Arbitrary-Rank Ablation (ARA)](https://github.com/p-e-w/heretic/pull/211) method
|
| 49 |
+
|
| 50 |
+
## Abliteration parameters
|
| 51 |
+
|
| 52 |
+
| Parameter | Value |
|
| 53 |
+
| :-------- | :---: |
|
| 54 |
+
| **start_layer_index** | 26 |
|
| 55 |
+
| **end_layer_index** | 46 |
|
| 56 |
+
| **preserve_good_behavior_weight** | 0.8239 |
|
| 57 |
+
| **steer_bad_behavior_weight** | 0.0001 |
|
| 58 |
+
| **overcorrect_relative_weight** | 1.1479 |
|
| 59 |
+
| **neighbor_count** | 10 |
|
| 60 |
+
|
| 61 |
+
## Targeted components
|
| 62 |
+
|
| 63 |
+
* attn.o_proj
|
| 64 |
+
|
| 65 |
+
## Performance
|
| 66 |
+
|
| 67 |
+
| Metric | This model | Original model ([Gemma4-Garnet-31B](https://huggingface.co/ConicCat/Gemma4-Garnet-31B)) |
|
| 68 |
+
| :----- | :--------: | :---------------------------: |
|
| 69 |
+
| **KL divergence** | <span style="color:darkgoldenrod">0.0368</span> | 0 *(by definition)* |
|
| 70 |
+
| **Refusals** | ✅ <span style="color:darkgreen">6/100</span> | ❌ <span style="color:blue">99/100</span> |
|
| 71 |
+
|
| 72 |
+
## PIQA test results:
|
| 73 |
+
|
| 74 |
+
<span style="color:blue">Original:</span>
|
| 75 |
+
|
| 76 |
+
- Total questions: 1838
|
| 77 |
+
- Correct: 1721
|
| 78 |
+
- Accuracy: 0.9363 (93.63%)
|
| 79 |
+
- Parse failures: 0
|
| 80 |
+
|
| 81 |
+
<span style="color:darkgreen">Heretic:</span>
|
| 82 |
+
|
| 83 |
+
- Total questions: 1838
|
| 84 |
+
- Correct: 1724
|
| 85 |
+
- Accuracy: 0.9380 (93.80%)
|
| 86 |
+
- Parse failures: 0
|
| 87 |
+
|
| 88 |
+
Lower refusals indicate fewer content restrictions, while lower KL divergence indicates more closeness to the original model's baseline. Higher refusals cause more rejections, objections, pushbacks, lecturing, censorship, softening and deflections. PIQA (Physical Intuition Question Answering) a ~1,800 questions tests common-sense understanding of how the physical world works with benchmark scores to measure physical reasoning ability.
|
| 89 |
+
|
| 90 |
+
## MMLU test results:
|
| 91 |
+
|
| 92 |
+
<span style="color:blue">Original:</span>
|
| 93 |
+
|
| 94 |
+
============================================================
|
| 95 |
+
|
| 96 |
+
- Total questions: 7021
|
| 97 |
+
|
| 98 |
+
- Correct: 6032
|
| 99 |
+
|
| 100 |
+
- Accuracy: 0.8591 (85.91%)
|
| 101 |
+
|
| 102 |
+
- Parse failures: 25
|
| 103 |
+
|
| 104 |
+
============================================================
|
| 105 |
+
|
| 106 |
+
Top subjects:
|
| 107 |
+
- professional_law: 0.7452 (585/785)
|
| 108 |
+
- moral_scenarios: 0.8167 (361/442)
|
| 109 |
+
- miscellaneous: 0.9217 (353/383)
|
| 110 |
+
- professional_psychology: 0.8987 (284/316)
|
| 111 |
+
- high_school_psychology: 0.9704 (262/270)
|
| 112 |
+
- high_school_macroeconomics: 0.9188 (181/197)
|
| 113 |
+
- prehistory: (157/172)
|
| 114 |
+
- moral_disputes: 0.8218 (143/174)
|
| 115 |
+
- elementary_mathematics: 0.9185 (169/184)
|
| 116 |
+
- philosophy: 0.8553 (141/159)
|
| 117 |
+
|
| 118 |
+
|
| 119 |
+
<span style="color:darkgreen">Heretic:</span>
|
| 120 |
+
|
| 121 |
+
============================================================
|
| 122 |
+
|
| 123 |
+
- Total questions: 7021
|
| 124 |
+
|
| 125 |
+
- Correct: 5954
|
| 126 |
+
|
| 127 |
+
- Accuracy: 0.8480 (84.80%)
|
| 128 |
+
|
| 129 |
+
- Parse failures: 21
|
| 130 |
+
|
| 131 |
+
============================================================
|
| 132 |
+
|
| 133 |
+
Top subjects:
|
| 134 |
+
- professional_law: 0.7223 (567/785)
|
| 135 |
+
- moral_scenarios: 0.7534 (333/442)
|
| 136 |
+
- miscellaneous: 0.9243 (354/383)
|
| 137 |
+
- professional_psychology: 0.8797 (278/316)
|
| 138 |
+
- high_school_psychology: 0.9667 (261/270)
|
| 139 |
+
- high_school_macroeconomics: 0.9137 (180/197)
|
| 140 |
+
- prehistory: 0.9186 (158/172)
|
| 141 |
+
- moral_disputes: 0.8103 (141/174)
|
| 142 |
+
- elementary_mathematics: 0.9239 (170/184)
|
| 143 |
+
- philosophy: 0.8239 (131/159)
|
| 144 |
+
|
| 145 |
+
MMLU - Massive Multitask Language Understanding, multiple-choice questions across 57 subjects (math, history, law, medicine, etc.).
|
| 146 |
+
|
| 147 |
+
-----
|
| 148 |
+
|
| 149 |
+
## Quantizations
|
| 150 |
+
|
| 151 |
+
| Filename | Quant | Description |
|
| 152 |
+
|----------|-------|-------------|
|
| 153 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-BF16.gguf | BF16 | Full precision |
|
| 154 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q8_0.gguf | Q8_0 | Near-lossless, recommended |
|
| 155 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q6_K.gguf | Q6_K | Excellent quality |
|
| 156 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q5_K_M.gguf | Q5_K_M | Good balance |
|
| 157 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q5_K_SQwen3.5-27B-ultra-uncensored-heretic-v2-v2-Q5_K_S.gguf | Q5_K_S | Smaller Q5 |
|
| 158 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q4_K_M.gguf | Q4_K_M | Good for limited VRAM |
|
| 159 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q4_K_S.gguf | Q4_K_S | Smaller Q4 |
|
| 160 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q3_K_L.gguf | Q3_K_L | Low VRAM, decent quality |
|
| 161 |
+
| Gemma-4-Garnet-31B-it-uncensored-heretic-Q3_K_M.gguf | Q3_K_M | Low VRAM, smaller |
|
| 162 |
+
|
| 163 |
+
## Vision Projector
|
| 164 |
+
|
| 165 |
+
| Filename | Quant | Description |
|
| 166 |
+
|----------|-------|-------------|
|
| 167 |
+
| Gemma-4-Garnet-31B-it-mmproj-BF16.gguf | BF16 | Native precision |
|
| 168 |
+
|
| 169 |
+
A Vision Projector File is Required for vision/multimodal capabilities. Use alongside any quantization above.
|
| 170 |
+
|
| 171 |
+
## Usage
|
| 172 |
+
|
| 173 |
+
Works with llama.cpp, LM Studio, Ollama, and other GGUF-compatible tools.
|
| 174 |
+
|
| 175 |
+
-----
|
| 176 |
+
|
| 177 |
+
# ConicCat/Gemma4-Garnet-31B
|
| 178 |
+
|
| 179 |
+
A finetune primarily focused on improving the prose and writing capabilities of Gemma 4. This does generalize strongly to roleplay and most other creative domains as well.
|
| 180 |
+
|
| 181 |
+
|
| 182 |
+
### Features:
|
| 183 |
+
* Improved longform writing capabilites; output context extension allows for prompting for up to 4000 words of text in one go.
|
| 184 |
+
* Markedly less AI slop and identifiable Gemini-isms in writing.
|
| 185 |
+
* Improved swipe or output diversity.
|
| 186 |
+
* Fewer 'soft' refusals in writing.
|
| 187 |
+
|
| 188 |
+
|
| 189 |
+
### Datasets
|
| 190 |
+
|
| 191 |
+
* internlm/Condor-SFT-20K for instruct; even though instruct capabilities are not the primary focus, adding some instruct data helps mitigate forgetting and maintains general intellect and instruction following capabilites.
|
| 192 |
+
* ConicCat/Gutenberg-SFT. A reformatted version of the original Gutenberg DPO dataset by jondurbin for SFT with some slight augmentation to address many of the samples being overly long.
|
| 193 |
+
* A dataset of backtranslated books. Unfortunately, I am unable to release this set as all of the data is under copyright.
|
| 194 |
+
* A dash of a certain third owned archive.
|