seungduk commited on
Commit
c7ac91f
·
verified ·
1 Parent(s): f87eb3e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,24 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf filter=lfs diff=lfs merge=lfs -text
37
+ IQ2_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_m.gguf filter=lfs diff=lfs merge=lfs -text
38
+ IQ2_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_s.gguf filter=lfs diff=lfs merge=lfs -text
39
+ IQ3_S/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf filter=lfs diff=lfs merge=lfs -text
40
+ IQ3_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf filter=lfs diff=lfs merge=lfs -text
41
+ IQ3_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xs.gguf filter=lfs diff=lfs merge=lfs -text
42
+ IQ3_XXS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xxs.gguf filter=lfs diff=lfs merge=lfs -text
43
+ IQ4_XS/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf filter=lfs diff=lfs merge=lfs -text
44
+ IQ4_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Q3_K_L/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Q3_K_L_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Q3_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Q3_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
49
+ Q5_K_M/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
50
+ Q5_K_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
51
+ Q5_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
52
+ Q5_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
53
+ Q6_K/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
54
+ Q6_K_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
55
+ Q8_0/YanoljaNEXT-Rosetta-4B-2510-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
56
+ imatrix.gguf filter=lfs diff=lfs merge=lfs -text
BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbb764230144b1821196a85153863bd6d1622186b7635863394261c5d1e3ff0d
3
+ size 7767803424
IQ2_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2d736fa59036a99d61ce88603a71498cd02787771e60fe395a7f0f23daee28f
3
+ size 1537982336
IQ2_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b00dd9d60cd6ce3de37f2c2b3e51e770219f73d6ae958d751f4b67e3fead4570
3
+ size 1449344896
IQ3_S/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:355c8ed24471dc5cf2831d24c9129b7e6481988c6018c5bcdc485a575923d480
3
+ size 1937363744
IQ3_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb07d75fe7ea5e76aef2e5d64091e750e8bb22bc327c7cc08b2b80734e715d8f
3
+ size 1937364096
IQ3_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84babc42c4df9fab82dd013d3fab7b29dfe1313f6bb4ccbc97c5a3e05fd33afa
3
+ size 1863390336
IQ3_XXS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xxs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f50df4eebe06688310932b72194cd608cb15156e38bdae327c36ff36c54d6543
3
+ size 1689452416
IQ4_XS/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a17126b851b4156f15a50843b9b021b326e937133e0af6b73af329e68ef7c52f
3
+ size 2279625504
IQ4_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30ad4723abfd3e3c7f6a841ea35ea2c40b0b16032c6daf0a72f1e22f2fa6ccdb
3
+ size 2263241856
Q3_K_L/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d57a282315d6038351a8366e1caca95e7fe80b9babb270ad92593ac7cbf040c5
3
+ size 2236085024
Q3_K_L_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe3207c95deab1ed3f61f82f4b070a1ce0b01c20f4734555c78b1b57a140e82b
3
+ size 2236085376
Q3_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7eec0d732e1abc65aea1505705c673548586f0db5c10ac1d77479f7ff209680
3
+ size 1937363744
Q3_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43bde6fead1473f8de196c881892c70006980004db2b270ed4721de9327612db
3
+ size 1937364096
Q5_K_M/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7ccede291346cc2554445dca82085ffcdd0af1caad0a942614dd3d12a394d49
3
+ size 2829697824
Q5_K_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:526eac30c887b237756d5ac19a2858b3cd4e27319f147a3132e339c1f813b411
3
+ size 2829698176
Q5_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ded15e8caf82e1540a11a830cfdb627ad90fa8a4c4887818eba4e56adcefb64a
3
+ size 2764591904
Q5_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:929ff041fa7e892f7e08ab3fa65b19428b993c2ae5ec85ec9e9d819cf261d4d9
3
+ size 2764592256
Q6_K/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:245dcd73830c8c9a42e18b6db0c892c012014ee4855306bf3df625869e9c82f5
3
+ size 3190739744
Q6_K_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4db88b92e2ed9e2c3308b66c851faf31fddc3e0e674c7d5cd0a0680eff76136b
3
+ size 3190740096
Q8_0/YanoljaNEXT-Rosetta-4B-2510-q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85f9be4da1b4be0214dcaf2b21c5adb5af503a749dc5a2db66511f1f179b541b
3
+ size 4130401824
README.md ADDED
@@ -0,0 +1,242 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ library_name: transformers
2
+ tags:
3
+ - translation
4
+ - llama-cpp
5
+ license: gemma
6
+ language:
7
+ - ar
8
+ - bg
9
+ - zh
10
+ - cs
11
+ - da
12
+ - nl
13
+ - en
14
+ - fi
15
+ - fr
16
+ - de
17
+ - el
18
+ - gu
19
+ - he
20
+ - hi
21
+ - hu
22
+ - id
23
+ - it
24
+ - ja
25
+ - ko
26
+ - fa
27
+ - pl
28
+ - pt
29
+ - ro
30
+ - ru
31
+ - sk
32
+ - es
33
+ - sv
34
+ - tl
35
+ - th
36
+ - tr
37
+ - uk
38
+ - vi
39
+ base_model: yanolja/YanoljaNEXT-Rosetta-12B-2510
40
+ ---
41
+
42
+ # YanoljaNEXT-Rosetta-4B-2510
43
+
44
+ <p style="text-align: center; margin: 0 auto 64px">
45
+ <img src="next_rosetta.png" style="width: 1096px">
46
+ </p>
47
+
48
+ This model is a fine-tuned version of [`google/gemma-3-4b-pt`](https://huggingface.co/google/gemma-3-4b-pt). As it is intended solely for text generation, we have extracted and utilized only the `Gemma3ForCausalLM` component from the original architecture.
49
+
50
+ Unlike our previous EEVE models, this model does not feature an expanded tokenizer.
51
+
52
+ - **Model Name:** `yanolja/YanoljaNEXT-Rosetta-4B-2510`
53
+ - **Base Model:** `google/gemma-3-4b-pt`
54
+
55
+ ## GGUF files
56
+
57
+ This folder contains ready-to-run GGUF files for llama.cpp.
58
+
59
+ - `BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf`: full-precision reference model
60
+ - Quantized variants (choose one based on your device and quality needs):
61
+ - K-family: `Q3_K_{S,L}`, `Q5_K_{S,M}`, `Q6_K`, `Q8_0`
62
+ - IQ-family: `IQ2_{S,M}`, `IQ3_{XXS,XS,S}`, `IQ4_{XS}`
63
+ - For many types there are matching `_IMX` folders. Files there were produced with an activation matrix (`imatrix.gguf`) and usually offer better quality at the same size. In this release, `IQ2_{S,M}` and `IQ3_{XXS,XS}` are IMX-only.
64
+
65
+ ## Model Description
66
+
67
+ This model is a 4-billion parameter, decoder-only language model built on the Gemma3 architecture and fine-tuned by Yanolja NEXT. It is specifically designed to translate structured data (JSON format) while preserving the original data structure.
68
+
69
+ The model was trained on a multilingual dataset covering the following languages equally:
70
+ - Arabic
71
+ - Bulgarian
72
+ - Chinese
73
+ - Czech
74
+ - Danish
75
+ - Dutch
76
+ - English
77
+ - Finnish
78
+ - French
79
+ - German
80
+ - Greek
81
+ - Gujarati
82
+ - Hebrew
83
+ - Hindi
84
+ - Hungarian
85
+ - Indonesian
86
+ - Italian
87
+ - Japanese
88
+ - Korean
89
+ - Persian
90
+ - Polish
91
+ - Portuguese
92
+ - Romanian
93
+ - Russian
94
+ - Slovak
95
+ - Spanish
96
+ - Swedish
97
+ - Tagalog
98
+ - Thai
99
+ - Turkish
100
+ - Ukrainian
101
+ - Vietnamese
102
+
103
+ While optimized for these languages, it may also perform effectively on other languages supported by the base Gemma3 model.
104
+
105
+ ## How to use
106
+
107
+ Use a recent build of `llama.cpp` that supports Gemma 3 models. Pick any GGUF file from this folder (a quantized variant is recommended for most users).
108
+
109
+ ```bash
110
+ # Example: use a Q5_K_M quantized file (adjust the path/model to your choice)
111
+ MODEL="path/to/YanoljaNEXT-Rosetta-4B-2510-q5_k_m.gguf"
112
+
113
+ # Build a formatted prompt using the included chat template roles
114
+ # (see release/YanoljaNEXT-Rosetta-4B-2510/chat_template.jinja)
115
+ read -r -d '' PROMPT <<'EOT'
116
+ <start_of_turn>instruction
117
+ Translate the user's text to Korean. Keep the JSON structure and keys.
118
+ Context: Simple introduction about a tech company.
119
+ Tone: Informative and helpful
120
+ Glossary:
121
+ - Yanolja NEXT -> 야놀자넥스트
122
+ - travel industry -> 여행 산업
123
+ Provide the final translation immediately without any other text.
124
+ <end_of_turn>
125
+ <start_of_turn>source
126
+ {"company_name": "Yanolja NEXT", "description": "Yanolja NEXT is a company that provides cutting-edge technology for the global travel industry."}
127
+ <end_of_turn>
128
+ <start_of_turn>translation\n
129
+ EOT
130
+
131
+ # Run llama.cpp (adjust -n/-c/--temp as needed)
132
+ llama-cli -m "$MODEL" -p "$PROMPT" -n 64 -c 4096 --temp 0.7 -no-cnv
133
+ ```
134
+
135
+ The model is optimized to output structured JSON for translations when appropriate.
136
+
137
+ ### REST server
138
+
139
+ ```bash
140
+ MODEL="path/to/YanoljaNEXT-Rosetta-4B-2510-q5_k_m.gguf"
141
+ llama-server -m "$MODEL" -c 4096 --host 0.0.0.0 --port 8080
142
+ ```
143
+
144
+ ### LM Studio / other GUIs
145
+
146
+ Import any of the `.gguf` files into your GUI of choice (LM Studio, KoboldCPP, text-generation-webui) and select chat mode. The embedded template in the GGUF will be used automatically by recent tools.
147
+
148
+ ## Training Procedure
149
+
150
+ ### Training Data
151
+ The translation datasets were synthesized using fineweb corpora.
152
+ - [FineWeb Edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu)
153
+ - [FineWeb2](https://huggingface.co/datasets/HuggingFaceFW/fineweb-2)
154
+
155
+ The model was fine-tuned on synthetic multilingual translation data to optimize performance across the supported language pairs.
156
+
157
+ ## Performance
158
+
159
+ ### Translation Quality Benchmarks
160
+
161
+ The following CHrF++ scores (WMT24++) demonstrate the model's competitive performance compared to other state-of-the-art translation models on English to Korean translation:
162
+
163
+ | Model | CHrF++ Score (WMT24++) |
164
+ |------------------------------------|--------------|
165
+ | google/gemini-2.5-flash-lite | 35.23 |
166
+ | **yanolja/YanoljaNEXT-Rosetta-4B-2510** | **35.09** |
167
+ | yanolja/YanoljaNEXT-Rosetta-12B | 34.75 |
168
+ | yanolja/YanoljaNEXT-Rosetta-20B | 33.87 |
169
+ | google/gemini-2.0-flash-001 | 33.81 |
170
+ | openai/gpt-oss-120b | 31.51 |
171
+ | **yanolja/YanoljaNEXT-Rosetta-4B** | **31.31** |
172
+ | openai/gpt-4.1-nano | 31.15 |
173
+ | Qwen/Qwen3-235B-A22B-Instruct-2507-FP8 | 31.02 |
174
+ | openai/gpt-oss-20b | 30.56 |
175
+ | google/gemma-3-27b-it | 30.05 |
176
+ | google/gemma-3-4b-pt | 27.53 |
177
+
178
+ YanoljaNEXT-Rosetta-4B-2510 achieves competitive translation quality while maintaining the efficiency of a 4B parameter model.
179
+ Scores for the other language pairs can be found in the [WMT24++ Evaluation Results](wmt24pp_12b.md).
180
+
181
+ ## Intended Uses & Limitations
182
+
183
+ This model is intended for translating structured data (JSON format) while preserving the original structure. It is particularly well-suited for tasks such as localizing product catalogs, translating hotel reviews, or handling any other structured content that requires accurate translation.
184
+
185
+ ### Limitations
186
+ The model is primarily optimized for processing JSON data.
187
+ Its performance on unstructured text or other data formats may vary.
188
+ In some cases, the model may produce invalid JSON, repetitive output, or inaccurate translations.
189
+
190
+ ### License
191
+ This model is released under the Gemma license, inherited from its base model, [`google/gemma-3-4b-pt`](https://huggingface.co/google/gemma-3-4b-pt). Please consult the official [Gemma license terms](https://ai.google.dev/gemma/terms) for detailed usage guidelines.
192
+
193
+ ## Acknowledgments
194
+ This work was supported by the Korea Creative Content Agency (KOCCA) grant, funded by the Ministry of Culture, Sports and Tourism (MCST) in 2025 (Project Name: _Cultivating Masters and Doctoral Experts to Lead Digital-Tech Tourism_, Project Number: RS-2024-00442006, Contribution Rate: 100%).
195
+
196
+ ## Citation
197
+
198
+ If you use this model, please consider citing:
199
+
200
+ ```
201
+ @misc{yanolja2025yanoljanextrosetta,
202
+ author = {Yanolja NEXT Co., Ltd.},
203
+ title = {YanoljaNEXT-Rosetta-4B-2510},
204
+ year = {2025},
205
+ publisher = {Hugging Face},
206
+ journal = {Hugging Face repository},
207
+ howpublished = {\\url{https://huggingface.co/yanolja/YanoljaNEXT-Rosetta-4B-2510}}
208
+ }
209
+ ```
210
+
211
+ ## References
212
+
213
+ This work utilizes several models and datasets. We would like to acknowledge the original authors for their valuable contributions to the field.
214
+
215
+ ```
216
+ @misc{gemma3,
217
+ author = {Google},
218
+ title = {Gemma 3},
219
+ year = {2024},
220
+ publisher = {Google DeepMind},
221
+ howpublished = {\\url{https://deepmind.google/models/gemma/gemma-3/}}
222
+ }
223
+
224
+ @misc{penedo2025fineweb2pipelinescale,
225
+ title = {FineWeb2: One Pipeline to Scale Them All -- Adapting Pre-Training Data Processing to Every Language},
226
+ author = {Guilherme Penedo and Hynek Kydlíček and Vinko Sabolčec and Bettina Messmer and Negar Foroutan and Amir Hossein Kargaran and Colin Raffel and Martin Jaggi and Leandro Von Werra and Thomas Wolf},
227
+ year = {2025},
228
+ eprint = {2506.20920},
229
+ archivePrefix = {arXiv},
230
+ primaryClass = {cs.CL},
231
+ url = {https://arxiv.org/abs/2506.20920},
232
+ }
233
+
234
+ @misc{lozhkov2024fineweb-edu,
235
+ author = {Lozhkov, Anton and Ben Allal, Loubna and von Werra, Leandro and Wolf, Thomas},
236
+ title = {FineWeb-Edu: the Finest Collection of Educational Content},
237
+ year = 2024,
238
+ url = {https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu},
239
+ doi = {10.57967/hf/2497},
240
+ publisher={Hugging Face}
241
+ }
242
+ ```
imatrix.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76e37ef58e1293da48ffb46003dad667d76e7159e1c750f8cb5b6e07f6611970
3
+ size 3448672