MuXodious's picture
Update README.md
982568e verified
|
raw
history blame
8.82 kB
---
tags:
- heretic
- uncensored
- decensored
- abliterated
base_model:
- Mistral-Nemo-Instruct-2407
- TheDrummer/Rocinante-XL-16B-v1
pipeline_tag: text-generation
---
This is a **Rocinante-XL-16B-v1** fine-tune, produced through P-E-W's [Heretic](https://github.com/p-e-w/heretic) (v1.2.0) abliteration engine with [Self-Organizing Maps & Magnitude-Preserving Orthogonal Ablation](https://github.com/p-e-w/heretic/pull/196) enabled.
---
<p>
<img src="https://img.shields.io/badge/HERESY_INDEX-ABSOLUTE-white?style=flat-square&labelColor=101010" align="right" width="250">
<b>Heretication Results</b>
<br clear="right">
<img src="https://img.shields.io/badge/RENEGADE_CHAPTER-SOMPOA-FCC900?style=flat-square&labelColor=101010" align="right" width="300">
</p>
<br clear="right">
| Score Metric | Value | Parameter | Value |
| :--- | :--- | :--- | :--- |
| **Refusals** | 3/416 | **direction_index** | 22.20 |
| **KL Divergence** | 0.0182 | **attn.o_proj.max_weights.0** | 0: 1.26 |
| **Initial Refusals** | 339/416 | **attn.o_proj.max_weights.1** | 1: 0.64 |
||| **attn.o_proj.max_weights.2** | 2: 1.41 |
||| **attn.o_proj.max_weights.3** | 3: 0.94 |
||| **attn.o_proj.max_weight_position** | 23.86 |
||| **attn.o_proj.min_weights.0** | 0: 0.97 |
||| **attn.o_proj.min_weights.1** | 1: 0.03 |
||| **attn.o_proj.min_weights.2** | 2: 1.18 |
||| **attn.o_proj.min_weights.3** | 3: 0.93 |
||| **attn.o_proj.min_weight_distance** | 18.57 |
||| **mlp.down_proj.max_weights.0** | 0: 1.23 |
||| **mlp.down_proj.max_weights.1** | 1: 0.70 |
||| **mlp.down_proj.max_weights.2** | 2: 1.35 |
||| **mlp.down_proj.max_weights.3** | 3: 0.86 |
||| **mlp.down_proj.max_weight_position** | 28.60 |
||| **mlp.down_proj.min_weights.0** | 0: 0.37 |
||| **mlp.down_proj.min_weights.1** | 1: 0.25 |
||| **mlp.down_proj.min_weights.2** | 2: 1.01 |
||| **mlp.down_proj.min_weights.3** | 3: 0.45 |
||| **mlp.down_proj.min_weight_distance** | 5.96 |
---
## Degree of Heretication
The **Heresy Index** weighs the resulting model's corruption by the process (KL Divergence & PIQA, Manual Response Eval) and its abolition of doctrine (Refusals) for a final verdict in classification.
| Index Entry | Classification | Analysis |
| :--- | :--- | :--- |
| ![Absolute](https://img.shields.io/badge/HERESY_INDEX-ABSOLUTE-white?style=flat-square&labelColor=101010) | **Absolute Heresy** | Near zero overt and secondary refusals with minimal to none model damage |
| ![Tainted](https://img.shields.io/badge/HERESY_INDEX-TAINTED-blueviolet?style=flat-square&labelColor=101010) | **Tainted Heresy** | Some residual secondary refusals and/or moderate model damage |
| ![Impotent](https://img.shields.io/badge/HERESY_INDEX-IMPOTENT-5c4033?style=flat-square&labelColor=101010) | **Impotent Heresy** | Lingering overt refusals and high model damage |
**Note**: This is an arbitrary and subjective classification inspired by Warhammer 40K, indended to serve as a signpost towards the model's performance.
---
**Appendix**
> Empty system prompt.
<details>
<summary>Heretication Rituals</summary>
```
Β» [Trial 93] Refusals: 3/416, KL divergence: 0.0182
[Trial 159] Refusals: 4/416, KL divergence: 0.0141
[Trial 80] Refusals: 9/416, KL divergence: 0.0140
[Trial 174] Refusals: 10/416, KL divergence: 0.0140
[Trial 163] Refusals: 12/416, KL divergence: 0.0132
[Trial 118] Refusals: 15/416, KL divergence: 0.0121
[Trial 82] Refusals: 18/416, KL divergence: 0.0099
[Trial 169] Refusals: 22/416, KL divergence: 0.0095
[Trial 119] Refusals: 35/416, KL divergence: 0.0091
[Trial 96] Refusals: 40/416, KL divergence: 0.0084
[Trial 100] Refusals: 45/416, KL divergence: 0.0067
[Trial 109] Refusals: 67/416, KL divergence: 0.0066
[Trial 62] Refusals: 155/416, KL divergence: 0.0065
[Trial 151] Refusals: 157/416, KL divergence: 0.0065
[Trial 164] Refusals: 168/416, KL divergence: 0.0060
[Trial 127] Refusals: 195/416, KL divergence: 0.0048
[Trial 139] Refusals: 263/416, KL divergence: 0.0041
[Trial 32] Refusals: 267/416, KL divergence: 0.0030
[Trial 101] Refusals: 313/416, KL divergence: 0.0016
[Trial 63] Refusals: 317/416, KL divergence: 0.0015
[Trial 181] Refusals: 330/416, KL divergence: 0.0014
[Trial 13] Refusals: 332/416, KL divergence: 0.0014
[Trial 59] Refusals: 333/416, KL divergence: 0.0011
[Trial 54] Refusals: 339/416, KL divergence: 0.0008
```
</details>
<details>
<summary>PIQA Benchmarks</summary>
```
┏━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━┓
┃ Benchmark ┃ Metric ┃ Value ┃
┑━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━┩
β”‚ PIQA Base β”‚ acc,none β”‚ 0.7900 β”‚
β”‚ β”‚ acc_stderr,none β”‚ 0.0095 β”‚
β”‚ β”‚ acc_norm,none β”‚ 0.8020 β”‚
β”‚ β”‚ acc_norm_stderr,none β”‚ 0.0093 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
┏━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━┓
┃ Benchmark ┃ Metric ┃ Value ┃
┑━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━┩
β”‚ PIQA T93 β”‚ acc,none β”‚ 0.7900 β”‚
β”‚ β”‚ acc_stderr,none β”‚ 0.0095 β”‚
β”‚ β”‚ acc_norm,none β”‚ 0.8030 β”‚
β”‚ β”‚ acc_norm_stderr,none β”‚ 0.0093 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
┏━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━┓
┃ Benchmark ┃ Metric ┃ Value ┃
┑━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━┩
β”‚ PIQA T159 β”‚ acc,none β”‚ 0.7878 β”‚
β”‚ β”‚ acc_stderr,none β”‚ 0.0095 β”‚
β”‚ β”‚ acc_norm,none β”‚ 0.8047 β”‚
β”‚ β”‚ acc_norm_stderr,none β”‚ 0.0092 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
┏━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━┓
┃ Benchmark ┃ Metric ┃ Value ┃
┑━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━┩
β”‚ PIQA T163 β”‚ acc,none β”‚ 0.7884 β”‚
β”‚ β”‚ acc_stderr,none β”‚ 0.0095 β”‚
β”‚ β”‚ acc_norm,none β”‚ 0.8036 β”‚
β”‚ β”‚ acc_norm_stderr,none β”‚ 0.0093 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
┏━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━┓
┃ Benchmark ┃ Metric ┃ Value ┃
┑━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━┩
β”‚ PIQA T80 β”‚ acc,none β”‚ 0.7884 β”‚
β”‚ β”‚ acc_stderr,none β”‚ 0.0095 β”‚
β”‚ β”‚ acc_norm,none β”‚ 0.8020 β”‚
β”‚ β”‚ acc_norm_stderr,none β”‚ 0.0093 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
┏━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━┓
┃ Benchmark ┃ Metric ┃ Value ┃
┑━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━┩
β”‚ PIQA T174 β”‚ acc,none β”‚ 0.7889 β”‚
β”‚ β”‚ acc_stderr,none β”‚ 0.0095 β”‚
β”‚ β”‚ acc_norm,none β”‚ 0.8014 β”‚
β”‚ β”‚ acc_norm_stderr,none β”‚ 0.0093 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
```
</details>
---
Mistral v3 Tekken or Metharme.
Can think via \<thinking\> or \<think\>
Just like Roci X but better.
(Model card still a WIP)
FP16: https://huggingface.co/TheDrummer/Rocinante-XL-16B-v1
GGUF: https://huggingface.co/TheDrummer/Rocinante-XL-16B-v1-GGUF