Commit ·
794847f
1
Parent(s): 396c5dd
Update README.md
Browse files
README.md
CHANGED
|
@@ -42,7 +42,7 @@ Just a fiction oriented 4bpw exl2 quantization of https://huggingface.co/jondurb
|
|
| 42 |
|
| 43 |
Quantized on 300K tokes of two Vicuna format chats, a sci fi story and a fiction story at a long context. This should yield better storywriting performance than the default exl2 quantization.
|
| 44 |
|
| 45 |
-
|
| 46 |
***
|
| 47 |
## Running
|
| 48 |
Being a Yi model, try running a lower temperature with ~0.05 MinP, a little repitition penalty, maybe mirostat with a low tau, and no other samplers. Yi tends to run "hot" by default.
|
|
|
|
| 42 |
|
| 43 |
Quantized on 300K tokes of two Vicuna format chats, a sci fi story and a fiction story at a long context. This should yield better storywriting performance than the default exl2 quantization.
|
| 44 |
|
| 45 |
+
If anyone wants sizes other than 4bpw, for more/less context or smaller GPUs, just ask.
|
| 46 |
***
|
| 47 |
## Running
|
| 48 |
Being a Yi model, try running a lower temperature with ~0.05 MinP, a little repitition penalty, maybe mirostat with a low tau, and no other samplers. Yi tends to run "hot" by default.
|