Update README.md
Browse files
README.md
CHANGED
|
@@ -13,7 +13,7 @@ tags:
|
|
| 13 |
library_name: transformers
|
| 14 |
---
|
| 15 |
|
| 16 |
-
# Light-
|
| 17 |
<a href="https://huggingface.co/qihoo360/Light-TLLM-7B" target="_blank" style="margin: 2px;">
|
| 18 |
<img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-FF6B6B" style="display: inline-block; vertical-align: middle;"/>
|
| 19 |
</a>
|
|
@@ -147,7 +147,7 @@ Light-TLLM-7B is evaluated on FLORES-Plus (90 directions) and standard instructi
|
|
| 147 |
| Seed-X-PPO-7B | Translation-focused | 25.9 | 22.6 | 10.5 | 13.3 | - | - | - | - |
|
| 148 |
| Hunyuan-MT-7B | Translation-focused | 24.6 | 23.4 | 14.8 | 16.6 | - | - | - | - |
|
| 149 |
| Light-TLLM-7B-SFT | Our models | 35.4 | 32.0 | 22.7 | 24.3 | 59.6 | 61.4 | 83.7 | 47.2 |
|
| 150 |
-
| **Light-TLLM-7B** | Our models | 36.1 | **32.7** | **23.1** | **24.9** | 60.9 | 63.2 | **85.2** | 48.5 |
|
| 151 |
|
| 152 |
- en->xx directions gain +1.1 BLEU over the next best 7B system while preserving reasoning accuracy (+1.3 MMLU over SFT).
|
| 153 |
- Average BLEU across all FLORES-Plus directions rises to 24.9 despite the compact 7B footprint.
|
|
@@ -173,7 +173,7 @@ RLVR introduces deterministic checks that reduce failure modes compared with gen
|
|
| 173 |
|
| 174 |
| Model | Language targeting | Length control | Format preservation | Code mixing | Overall |
|
| 175 |
| --- | --- | --- | --- | --- | --- |
|
| 176 |
-
| **Light-TLLM-7B** | **97.8** | 99.2 | **92.15** | 92.3 | **95.3** |
|
| 177 |
| Qwen2.5-7B-Instruct | 92.0 | 97.0 | 51.8 | 62.8 | 75.9 |
|
| 178 |
| Gemma3-27B-IT | 97.4 | 91.6 | 42.1 | 90.9 | 80.5 |
|
| 179 |
| Qwen-MT-Plus | 97.6 | **99.8** | 82.5 | 94.8 | 93.6 |
|
|
|
|
| 13 |
library_name: transformers
|
| 14 |
---
|
| 15 |
|
| 16 |
+
# Light-TLLM-7B
|
| 17 |
<a href="https://huggingface.co/qihoo360/Light-TLLM-7B" target="_blank" style="margin: 2px;">
|
| 18 |
<img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-FF6B6B" style="display: inline-block; vertical-align: middle;"/>
|
| 19 |
</a>
|
|
|
|
| 147 |
| Seed-X-PPO-7B | Translation-focused | 25.9 | 22.6 | 10.5 | 13.3 | - | - | - | - |
|
| 148 |
| Hunyuan-MT-7B | Translation-focused | 24.6 | 23.4 | 14.8 | 16.6 | - | - | - | - |
|
| 149 |
| Light-TLLM-7B-SFT | Our models | 35.4 | 32.0 | 22.7 | 24.3 | 59.6 | 61.4 | 83.7 | 47.2 |
|
| 150 |
+
| **Light-TLLM-7B-RL** | Our models | 36.1 | **32.7** | **23.1** | **24.9** | 60.9 | 63.2 | **85.2** | 48.5 |
|
| 151 |
|
| 152 |
- en->xx directions gain +1.1 BLEU over the next best 7B system while preserving reasoning accuracy (+1.3 MMLU over SFT).
|
| 153 |
- Average BLEU across all FLORES-Plus directions rises to 24.9 despite the compact 7B footprint.
|
|
|
|
| 173 |
|
| 174 |
| Model | Language targeting | Length control | Format preservation | Code mixing | Overall |
|
| 175 |
| --- | --- | --- | --- | --- | --- |
|
| 176 |
+
| **Light-TLLM-7B-RL** | **97.8** | 99.2 | **92.15** | 92.3 | **95.3** |
|
| 177 |
| Qwen2.5-7B-Instruct | 92.0 | 97.0 | 51.8 | 62.8 | 75.9 |
|
| 178 |
| Gemma3-27B-IT | 97.4 | 91.6 | 42.1 | 90.9 | 80.5 |
|
| 179 |
| Qwen-MT-Plus | 97.6 | **99.8** | 82.5 | 94.8 | 93.6 |
|