Upload folder using huggingface_hub
Browse files- .gitattributes +21 -0
- BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf +3 -0
- IQ2_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_m.gguf +3 -0
- IQ2_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_s.gguf +3 -0
- IQ3_S/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf +3 -0
- IQ3_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf +3 -0
- IQ3_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xs.gguf +3 -0
- IQ3_XXS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xxs.gguf +3 -0
- IQ4_XS/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf +3 -0
- IQ4_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf +3 -0
- Q3_K_L/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf +3 -0
- Q3_K_L_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf +3 -0
- Q3_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf +3 -0
- Q3_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf +3 -0
- Q5_K_M/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf +3 -0
- Q5_K_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf +3 -0
- Q5_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf +3 -0
- Q5_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf +3 -0
- Q6_K/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf +3 -0
- Q6_K_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf +3 -0
- Q8_0/YanoljaNEXT-Rosetta-4B-2510-q8_0.gguf +3 -0
- README.md +242 -0
- imatrix.gguf +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,24 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
IQ2_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
IQ2_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
IQ3_S/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
IQ3_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
IQ3_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xs.gguf filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
IQ3_XXS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xxs.gguf filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
IQ4_XS/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
IQ4_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
Q3_K_L/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
Q3_K_L_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
Q3_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
Q3_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
Q5_K_M/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
Q5_K_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text
|
| 51 |
+
Q5_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
Q5_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
Q6_K/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
|
| 54 |
+
Q6_K_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
Q8_0/YanoljaNEXT-Rosetta-4B-2510-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
imatrix.gguf filter=lfs diff=lfs merge=lfs -text
|
BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cbb764230144b1821196a85153863bd6d1622186b7635863394261c5d1e3ff0d
|
| 3 |
+
size 7767803424
|
IQ2_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_m.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a2d736fa59036a99d61ce88603a71498cd02787771e60fe395a7f0f23daee28f
|
| 3 |
+
size 1537982336
|
IQ2_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq2_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b00dd9d60cd6ce3de37f2c2b3e51e770219f73d6ae958d751f4b67e3fead4570
|
| 3 |
+
size 1449344896
|
IQ3_S/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:355c8ed24471dc5cf2831d24c9129b7e6481988c6018c5bcdc485a575923d480
|
| 3 |
+
size 1937363744
|
IQ3_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cb07d75fe7ea5e76aef2e5d64091e750e8bb22bc327c7cc08b2b80734e715d8f
|
| 3 |
+
size 1937364096
|
IQ3_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xs.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:84babc42c4df9fab82dd013d3fab7b29dfe1313f6bb4ccbc97c5a3e05fd33afa
|
| 3 |
+
size 1863390336
|
IQ3_XXS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq3_xxs.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f50df4eebe06688310932b72194cd608cb15156e38bdae327c36ff36c54d6543
|
| 3 |
+
size 1689452416
|
IQ4_XS/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a17126b851b4156f15a50843b9b021b326e937133e0af6b73af329e68ef7c52f
|
| 3 |
+
size 2279625504
|
IQ4_XS_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-iq4_xs.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:30ad4723abfd3e3c7f6a841ea35ea2c40b0b16032c6daf0a72f1e22f2fa6ccdb
|
| 3 |
+
size 2263241856
|
Q3_K_L/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d57a282315d6038351a8366e1caca95e7fe80b9babb270ad92593ac7cbf040c5
|
| 3 |
+
size 2236085024
|
Q3_K_L_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_l.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe3207c95deab1ed3f61f82f4b070a1ce0b01c20f4734555c78b1b57a140e82b
|
| 3 |
+
size 2236085376
|
Q3_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a7eec0d732e1abc65aea1505705c673548586f0db5c10ac1d77479f7ff209680
|
| 3 |
+
size 1937363744
|
Q3_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q3_k_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:43bde6fead1473f8de196c881892c70006980004db2b270ed4721de9327612db
|
| 3 |
+
size 1937364096
|
Q5_K_M/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d7ccede291346cc2554445dca82085ffcdd0af1caad0a942614dd3d12a394d49
|
| 3 |
+
size 2829697824
|
Q5_K_M_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_m.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:526eac30c887b237756d5ac19a2858b3cd4e27319f147a3132e339c1f813b411
|
| 3 |
+
size 2829698176
|
Q5_K_S/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ded15e8caf82e1540a11a830cfdb627ad90fa8a4c4887818eba4e56adcefb64a
|
| 3 |
+
size 2764591904
|
Q5_K_S_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q5_k_s.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:929ff041fa7e892f7e08ab3fa65b19428b993c2ae5ec85ec9e9d819cf261d4d9
|
| 3 |
+
size 2764592256
|
Q6_K/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:245dcd73830c8c9a42e18b6db0c892c012014ee4855306bf3df625869e9c82f5
|
| 3 |
+
size 3190739744
|
Q6_K_IMX/YanoljaNEXT-Rosetta-4B-2510-bf16-q6_k.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4db88b92e2ed9e2c3308b66c851faf31fddc3e0e674c7d5cd0a0680eff76136b
|
| 3 |
+
size 3190740096
|
Q8_0/YanoljaNEXT-Rosetta-4B-2510-q8_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:85f9be4da1b4be0214dcaf2b21c5adb5af503a749dc5a2db66511f1f179b541b
|
| 3 |
+
size 4130401824
|
README.md
ADDED
|
@@ -0,0 +1,242 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
library_name: transformers
|
| 2 |
+
tags:
|
| 3 |
+
- translation
|
| 4 |
+
- llama-cpp
|
| 5 |
+
license: gemma
|
| 6 |
+
language:
|
| 7 |
+
- ar
|
| 8 |
+
- bg
|
| 9 |
+
- zh
|
| 10 |
+
- cs
|
| 11 |
+
- da
|
| 12 |
+
- nl
|
| 13 |
+
- en
|
| 14 |
+
- fi
|
| 15 |
+
- fr
|
| 16 |
+
- de
|
| 17 |
+
- el
|
| 18 |
+
- gu
|
| 19 |
+
- he
|
| 20 |
+
- hi
|
| 21 |
+
- hu
|
| 22 |
+
- id
|
| 23 |
+
- it
|
| 24 |
+
- ja
|
| 25 |
+
- ko
|
| 26 |
+
- fa
|
| 27 |
+
- pl
|
| 28 |
+
- pt
|
| 29 |
+
- ro
|
| 30 |
+
- ru
|
| 31 |
+
- sk
|
| 32 |
+
- es
|
| 33 |
+
- sv
|
| 34 |
+
- tl
|
| 35 |
+
- th
|
| 36 |
+
- tr
|
| 37 |
+
- uk
|
| 38 |
+
- vi
|
| 39 |
+
base_model: yanolja/YanoljaNEXT-Rosetta-12B-2510
|
| 40 |
+
---
|
| 41 |
+
|
| 42 |
+
# YanoljaNEXT-Rosetta-4B-2510
|
| 43 |
+
|
| 44 |
+
<p style="text-align: center; margin: 0 auto 64px">
|
| 45 |
+
<img src="next_rosetta.png" style="width: 1096px">
|
| 46 |
+
</p>
|
| 47 |
+
|
| 48 |
+
This model is a fine-tuned version of [`google/gemma-3-4b-pt`](https://huggingface.co/google/gemma-3-4b-pt). As it is intended solely for text generation, we have extracted and utilized only the `Gemma3ForCausalLM` component from the original architecture.
|
| 49 |
+
|
| 50 |
+
Unlike our previous EEVE models, this model does not feature an expanded tokenizer.
|
| 51 |
+
|
| 52 |
+
- **Model Name:** `yanolja/YanoljaNEXT-Rosetta-4B-2510`
|
| 53 |
+
- **Base Model:** `google/gemma-3-4b-pt`
|
| 54 |
+
|
| 55 |
+
## GGUF files
|
| 56 |
+
|
| 57 |
+
This folder contains ready-to-run GGUF files for llama.cpp.
|
| 58 |
+
|
| 59 |
+
- `BF16/YanoljaNEXT-Rosetta-4B-2510-bf16.gguf`: full-precision reference model
|
| 60 |
+
- Quantized variants (choose one based on your device and quality needs):
|
| 61 |
+
- K-family: `Q3_K_{S,L}`, `Q5_K_{S,M}`, `Q6_K`, `Q8_0`
|
| 62 |
+
- IQ-family: `IQ2_{S,M}`, `IQ3_{XXS,XS,S}`, `IQ4_{XS}`
|
| 63 |
+
- For many types there are matching `_IMX` folders. Files there were produced with an activation matrix (`imatrix.gguf`) and usually offer better quality at the same size. In this release, `IQ2_{S,M}` and `IQ3_{XXS,XS}` are IMX-only.
|
| 64 |
+
|
| 65 |
+
## Model Description
|
| 66 |
+
|
| 67 |
+
This model is a 4-billion parameter, decoder-only language model built on the Gemma3 architecture and fine-tuned by Yanolja NEXT. It is specifically designed to translate structured data (JSON format) while preserving the original data structure.
|
| 68 |
+
|
| 69 |
+
The model was trained on a multilingual dataset covering the following languages equally:
|
| 70 |
+
- Arabic
|
| 71 |
+
- Bulgarian
|
| 72 |
+
- Chinese
|
| 73 |
+
- Czech
|
| 74 |
+
- Danish
|
| 75 |
+
- Dutch
|
| 76 |
+
- English
|
| 77 |
+
- Finnish
|
| 78 |
+
- French
|
| 79 |
+
- German
|
| 80 |
+
- Greek
|
| 81 |
+
- Gujarati
|
| 82 |
+
- Hebrew
|
| 83 |
+
- Hindi
|
| 84 |
+
- Hungarian
|
| 85 |
+
- Indonesian
|
| 86 |
+
- Italian
|
| 87 |
+
- Japanese
|
| 88 |
+
- Korean
|
| 89 |
+
- Persian
|
| 90 |
+
- Polish
|
| 91 |
+
- Portuguese
|
| 92 |
+
- Romanian
|
| 93 |
+
- Russian
|
| 94 |
+
- Slovak
|
| 95 |
+
- Spanish
|
| 96 |
+
- Swedish
|
| 97 |
+
- Tagalog
|
| 98 |
+
- Thai
|
| 99 |
+
- Turkish
|
| 100 |
+
- Ukrainian
|
| 101 |
+
- Vietnamese
|
| 102 |
+
|
| 103 |
+
While optimized for these languages, it may also perform effectively on other languages supported by the base Gemma3 model.
|
| 104 |
+
|
| 105 |
+
## How to use
|
| 106 |
+
|
| 107 |
+
Use a recent build of `llama.cpp` that supports Gemma 3 models. Pick any GGUF file from this folder (a quantized variant is recommended for most users).
|
| 108 |
+
|
| 109 |
+
```bash
|
| 110 |
+
# Example: use a Q5_K_M quantized file (adjust the path/model to your choice)
|
| 111 |
+
MODEL="path/to/YanoljaNEXT-Rosetta-4B-2510-q5_k_m.gguf"
|
| 112 |
+
|
| 113 |
+
# Build a formatted prompt using the included chat template roles
|
| 114 |
+
# (see release/YanoljaNEXT-Rosetta-4B-2510/chat_template.jinja)
|
| 115 |
+
read -r -d '' PROMPT <<'EOT'
|
| 116 |
+
<start_of_turn>instruction
|
| 117 |
+
Translate the user's text to Korean. Keep the JSON structure and keys.
|
| 118 |
+
Context: Simple introduction about a tech company.
|
| 119 |
+
Tone: Informative and helpful
|
| 120 |
+
Glossary:
|
| 121 |
+
- Yanolja NEXT -> 야놀자넥스트
|
| 122 |
+
- travel industry -> 여행 산업
|
| 123 |
+
Provide the final translation immediately without any other text.
|
| 124 |
+
<end_of_turn>
|
| 125 |
+
<start_of_turn>source
|
| 126 |
+
{"company_name": "Yanolja NEXT", "description": "Yanolja NEXT is a company that provides cutting-edge technology for the global travel industry."}
|
| 127 |
+
<end_of_turn>
|
| 128 |
+
<start_of_turn>translation\n
|
| 129 |
+
EOT
|
| 130 |
+
|
| 131 |
+
# Run llama.cpp (adjust -n/-c/--temp as needed)
|
| 132 |
+
llama-cli -m "$MODEL" -p "$PROMPT" -n 64 -c 4096 --temp 0.7 -no-cnv
|
| 133 |
+
```
|
| 134 |
+
|
| 135 |
+
The model is optimized to output structured JSON for translations when appropriate.
|
| 136 |
+
|
| 137 |
+
### REST server
|
| 138 |
+
|
| 139 |
+
```bash
|
| 140 |
+
MODEL="path/to/YanoljaNEXT-Rosetta-4B-2510-q5_k_m.gguf"
|
| 141 |
+
llama-server -m "$MODEL" -c 4096 --host 0.0.0.0 --port 8080
|
| 142 |
+
```
|
| 143 |
+
|
| 144 |
+
### LM Studio / other GUIs
|
| 145 |
+
|
| 146 |
+
Import any of the `.gguf` files into your GUI of choice (LM Studio, KoboldCPP, text-generation-webui) and select chat mode. The embedded template in the GGUF will be used automatically by recent tools.
|
| 147 |
+
|
| 148 |
+
## Training Procedure
|
| 149 |
+
|
| 150 |
+
### Training Data
|
| 151 |
+
The translation datasets were synthesized using fineweb corpora.
|
| 152 |
+
- [FineWeb Edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu)
|
| 153 |
+
- [FineWeb2](https://huggingface.co/datasets/HuggingFaceFW/fineweb-2)
|
| 154 |
+
|
| 155 |
+
The model was fine-tuned on synthetic multilingual translation data to optimize performance across the supported language pairs.
|
| 156 |
+
|
| 157 |
+
## Performance
|
| 158 |
+
|
| 159 |
+
### Translation Quality Benchmarks
|
| 160 |
+
|
| 161 |
+
The following CHrF++ scores (WMT24++) demonstrate the model's competitive performance compared to other state-of-the-art translation models on English to Korean translation:
|
| 162 |
+
|
| 163 |
+
| Model | CHrF++ Score (WMT24++) |
|
| 164 |
+
|------------------------------------|--------------|
|
| 165 |
+
| google/gemini-2.5-flash-lite | 35.23 |
|
| 166 |
+
| **yanolja/YanoljaNEXT-Rosetta-4B-2510** | **35.09** |
|
| 167 |
+
| yanolja/YanoljaNEXT-Rosetta-12B | 34.75 |
|
| 168 |
+
| yanolja/YanoljaNEXT-Rosetta-20B | 33.87 |
|
| 169 |
+
| google/gemini-2.0-flash-001 | 33.81 |
|
| 170 |
+
| openai/gpt-oss-120b | 31.51 |
|
| 171 |
+
| **yanolja/YanoljaNEXT-Rosetta-4B** | **31.31** |
|
| 172 |
+
| openai/gpt-4.1-nano | 31.15 |
|
| 173 |
+
| Qwen/Qwen3-235B-A22B-Instruct-2507-FP8 | 31.02 |
|
| 174 |
+
| openai/gpt-oss-20b | 30.56 |
|
| 175 |
+
| google/gemma-3-27b-it | 30.05 |
|
| 176 |
+
| google/gemma-3-4b-pt | 27.53 |
|
| 177 |
+
|
| 178 |
+
YanoljaNEXT-Rosetta-4B-2510 achieves competitive translation quality while maintaining the efficiency of a 4B parameter model.
|
| 179 |
+
Scores for the other language pairs can be found in the [WMT24++ Evaluation Results](wmt24pp_12b.md).
|
| 180 |
+
|
| 181 |
+
## Intended Uses & Limitations
|
| 182 |
+
|
| 183 |
+
This model is intended for translating structured data (JSON format) while preserving the original structure. It is particularly well-suited for tasks such as localizing product catalogs, translating hotel reviews, or handling any other structured content that requires accurate translation.
|
| 184 |
+
|
| 185 |
+
### Limitations
|
| 186 |
+
The model is primarily optimized for processing JSON data.
|
| 187 |
+
Its performance on unstructured text or other data formats may vary.
|
| 188 |
+
In some cases, the model may produce invalid JSON, repetitive output, or inaccurate translations.
|
| 189 |
+
|
| 190 |
+
### License
|
| 191 |
+
This model is released under the Gemma license, inherited from its base model, [`google/gemma-3-4b-pt`](https://huggingface.co/google/gemma-3-4b-pt). Please consult the official [Gemma license terms](https://ai.google.dev/gemma/terms) for detailed usage guidelines.
|
| 192 |
+
|
| 193 |
+
## Acknowledgments
|
| 194 |
+
This work was supported by the Korea Creative Content Agency (KOCCA) grant, funded by the Ministry of Culture, Sports and Tourism (MCST) in 2025 (Project Name: _Cultivating Masters and Doctoral Experts to Lead Digital-Tech Tourism_, Project Number: RS-2024-00442006, Contribution Rate: 100%).
|
| 195 |
+
|
| 196 |
+
## Citation
|
| 197 |
+
|
| 198 |
+
If you use this model, please consider citing:
|
| 199 |
+
|
| 200 |
+
```
|
| 201 |
+
@misc{yanolja2025yanoljanextrosetta,
|
| 202 |
+
author = {Yanolja NEXT Co., Ltd.},
|
| 203 |
+
title = {YanoljaNEXT-Rosetta-4B-2510},
|
| 204 |
+
year = {2025},
|
| 205 |
+
publisher = {Hugging Face},
|
| 206 |
+
journal = {Hugging Face repository},
|
| 207 |
+
howpublished = {\\url{https://huggingface.co/yanolja/YanoljaNEXT-Rosetta-4B-2510}}
|
| 208 |
+
}
|
| 209 |
+
```
|
| 210 |
+
|
| 211 |
+
## References
|
| 212 |
+
|
| 213 |
+
This work utilizes several models and datasets. We would like to acknowledge the original authors for their valuable contributions to the field.
|
| 214 |
+
|
| 215 |
+
```
|
| 216 |
+
@misc{gemma3,
|
| 217 |
+
author = {Google},
|
| 218 |
+
title = {Gemma 3},
|
| 219 |
+
year = {2024},
|
| 220 |
+
publisher = {Google DeepMind},
|
| 221 |
+
howpublished = {\\url{https://deepmind.google/models/gemma/gemma-3/}}
|
| 222 |
+
}
|
| 223 |
+
|
| 224 |
+
@misc{penedo2025fineweb2pipelinescale,
|
| 225 |
+
title = {FineWeb2: One Pipeline to Scale Them All -- Adapting Pre-Training Data Processing to Every Language},
|
| 226 |
+
author = {Guilherme Penedo and Hynek Kydlíček and Vinko Sabolčec and Bettina Messmer and Negar Foroutan and Amir Hossein Kargaran and Colin Raffel and Martin Jaggi and Leandro Von Werra and Thomas Wolf},
|
| 227 |
+
year = {2025},
|
| 228 |
+
eprint = {2506.20920},
|
| 229 |
+
archivePrefix = {arXiv},
|
| 230 |
+
primaryClass = {cs.CL},
|
| 231 |
+
url = {https://arxiv.org/abs/2506.20920},
|
| 232 |
+
}
|
| 233 |
+
|
| 234 |
+
@misc{lozhkov2024fineweb-edu,
|
| 235 |
+
author = {Lozhkov, Anton and Ben Allal, Loubna and von Werra, Leandro and Wolf, Thomas},
|
| 236 |
+
title = {FineWeb-Edu: the Finest Collection of Educational Content},
|
| 237 |
+
year = 2024,
|
| 238 |
+
url = {https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu},
|
| 239 |
+
doi = {10.57967/hf/2497},
|
| 240 |
+
publisher={Hugging Face}
|
| 241 |
+
}
|
| 242 |
+
```
|
imatrix.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76e37ef58e1293da48ffb46003dad667d76e7159e1c750f8cb5b6e07f6611970
|
| 3 |
+
size 3448672
|