HailJebus commited on
Commit
16dfb67
·
verified ·
1 Parent(s): c05c77d

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +88 -0
README.md ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ thumbnail: https://cdn-uploads.huggingface.co/production/uploads/67c10cfba43d7939d60160ff/Cjyto1cPNAwK2f_-uMyLz.png
3
+ language:
4
+ - en
5
+ license: other
6
+ inference: false
7
+ tags:
8
+ - nsfw
9
+ - explicit
10
+ - roleplay
11
+ - mixed-AI
12
+ - furry
13
+ - anthro
14
+ - dark
15
+ - chat
16
+ - llama-cpp
17
+ - gguf-my-repo
18
+ pipeline_tag: text-generation
19
+ library_name: transformers
20
+ base_model: Mawdistical/Kuwutu-7B
21
+ datasets:
22
+ - Delta-Vector/Hydrus-General-Reasoning
23
+ - Delta-Vector/Hydrus-IF-Mix-Ai2
24
+ - Delta-Vector/Hydrus-Army-Inst
25
+ - Delta-Vector/Hydrus-AM-thinking-Science
26
+ - Delta-Vector/Hydrus-AM-Thinking-Code-Filtered
27
+ - Delta-Vector/Hydrus-AM-Thinking-IF-No-Think
28
+ - Delta-Vector/Hydrus-Tulu-SFT-Mix-V2
29
+ - Delta-Vector/Hydrus-System-Chat-2.0
30
+ - Delta-Vector/Orion-Praxis-Co-Writer
31
+ - Delta-Vector/Orion-Co-Writer-51K
32
+ - Delta-Vector/Orion-Creative_Writing-Complexity
33
+ - Delta-Vector/Orion-vanilla-backrooms-claude-sharegpt
34
+ - Delta-Vector/Hydrus-AM-Thinking-Multi-Turn
35
+ - PocketDoc/Dans-Failuremaxx-Adventure
36
+ - PocketDoc/Dans-Logicmaxx-SAT-AP
37
+ - PocketDoc/Dans-MemoryCore-CoreCurriculum-Small
38
+ - PocketDoc/Dans-Taskmaxx-DataPrepper
39
+ - PocketDoc/Dans-Prosemaxx-Instructwriter-Long
40
+ - PocketDoc/Dans-Prosemaxx-InstructWriter-ZeroShot-2
41
+ - PocketDoc/Dans-Prosemaxx-InstructWriter-ZeroShot-3
42
+ - PocketDoc/Dans-Prosemaxx-InstructWriter-Continue-2
43
+ - PocketDoc/Dans-Systemmaxx
44
+ ---
45
+
46
+ # HailJebus/Kuwutu-7B-Q4_0-GGUF
47
+ This model was converted to GGUF format from [`Mawdistical/Kuwutu-7B`](https://huggingface.co/Mawdistical/Kuwutu-7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
48
+ Refer to the [original model card](https://huggingface.co/Mawdistical/Kuwutu-7B) for more details on the model.
49
+
50
+ ## Use with llama.cpp
51
+ Install llama.cpp through brew (works on Mac and Linux)
52
+
53
+ ```bash
54
+ brew install llama.cpp
55
+
56
+ ```
57
+ Invoke the llama.cpp server or the CLI.
58
+
59
+ ### CLI:
60
+ ```bash
61
+ llama-cli --hf-repo HailJebus/Kuwutu-7B-Q4_0-GGUF --hf-file kuwutu-7b-q4_0.gguf -p "The meaning to life and the universe is"
62
+ ```
63
+
64
+ ### Server:
65
+ ```bash
66
+ llama-server --hf-repo HailJebus/Kuwutu-7B-Q4_0-GGUF --hf-file kuwutu-7b-q4_0.gguf -c 2048
67
+ ```
68
+
69
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
70
+
71
+ Step 1: Clone llama.cpp from GitHub.
72
+ ```
73
+ git clone https://github.com/ggerganov/llama.cpp
74
+ ```
75
+
76
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
77
+ ```
78
+ cd llama.cpp && LLAMA_CURL=1 make
79
+ ```
80
+
81
+ Step 3: Run inference through the main binary.
82
+ ```
83
+ ./llama-cli --hf-repo HailJebus/Kuwutu-7B-Q4_0-GGUF --hf-file kuwutu-7b-q4_0.gguf -p "The meaning to life and the universe is"
84
+ ```
85
+ or
86
+ ```
87
+ ./llama-server --hf-repo HailJebus/Kuwutu-7B-Q4_0-GGUF --hf-file kuwutu-7b-q4_0.gguf -c 2048
88
+ ```