khazic commited on
Commit
ee51d21
·
verified ·
1 Parent(s): 4e3cb4a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -13,7 +13,7 @@ tags:
13
  library_name: transformers
14
  ---
15
 
16
- # Light-MT-7B
17
  <a href="https://huggingface.co/qihoo360/Light-TLLM-7B" target="_blank" style="margin: 2px;">
18
  <img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-FF6B6B" style="display: inline-block; vertical-align: middle;"/>
19
  </a>
@@ -147,7 +147,7 @@ Light-TLLM-7B is evaluated on FLORES-Plus (90 directions) and standard instructi
147
  | Seed-X-PPO-7B | Translation-focused | 25.9 | 22.6 | 10.5 | 13.3 | - | - | - | - |
148
  | Hunyuan-MT-7B | Translation-focused | 24.6 | 23.4 | 14.8 | 16.6 | - | - | - | - |
149
  | Light-TLLM-7B-SFT | Our models | 35.4 | 32.0 | 22.7 | 24.3 | 59.6 | 61.4 | 83.7 | 47.2 |
150
- | **Light-TLLM-7B** | Our models | 36.1 | **32.7** | **23.1** | **24.9** | 60.9 | 63.2 | **85.2** | 48.5 |
151
 
152
  - en->xx directions gain +1.1 BLEU over the next best 7B system while preserving reasoning accuracy (+1.3 MMLU over SFT).
153
  - Average BLEU across all FLORES-Plus directions rises to 24.9 despite the compact 7B footprint.
@@ -173,7 +173,7 @@ RLVR introduces deterministic checks that reduce failure modes compared with gen
173
 
174
  | Model | Language targeting | Length control | Format preservation | Code mixing | Overall |
175
  | --- | --- | --- | --- | --- | --- |
176
- | **Light-TLLM-7B** | **97.8** | 99.2 | **92.15** | 92.3 | **95.3** |
177
  | Qwen2.5-7B-Instruct | 92.0 | 97.0 | 51.8 | 62.8 | 75.9 |
178
  | Gemma3-27B-IT | 97.4 | 91.6 | 42.1 | 90.9 | 80.5 |
179
  | Qwen-MT-Plus | 97.6 | **99.8** | 82.5 | 94.8 | 93.6 |
 
13
  library_name: transformers
14
  ---
15
 
16
+ # Light-TLLM-7B
17
  <a href="https://huggingface.co/qihoo360/Light-TLLM-7B" target="_blank" style="margin: 2px;">
18
  <img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-FF6B6B" style="display: inline-block; vertical-align: middle;"/>
19
  </a>
 
147
  | Seed-X-PPO-7B | Translation-focused | 25.9 | 22.6 | 10.5 | 13.3 | - | - | - | - |
148
  | Hunyuan-MT-7B | Translation-focused | 24.6 | 23.4 | 14.8 | 16.6 | - | - | - | - |
149
  | Light-TLLM-7B-SFT | Our models | 35.4 | 32.0 | 22.7 | 24.3 | 59.6 | 61.4 | 83.7 | 47.2 |
150
+ | **Light-TLLM-7B-RL** | Our models | 36.1 | **32.7** | **23.1** | **24.9** | 60.9 | 63.2 | **85.2** | 48.5 |
151
 
152
  - en->xx directions gain +1.1 BLEU over the next best 7B system while preserving reasoning accuracy (+1.3 MMLU over SFT).
153
  - Average BLEU across all FLORES-Plus directions rises to 24.9 despite the compact 7B footprint.
 
173
 
174
  | Model | Language targeting | Length control | Format preservation | Code mixing | Overall |
175
  | --- | --- | --- | --- | --- | --- |
176
+ | **Light-TLLM-7B-RL** | **97.8** | 99.2 | **92.15** | 92.3 | **95.3** |
177
  | Qwen2.5-7B-Instruct | 92.0 | 97.0 | 51.8 | 62.8 | 75.9 |
178
  | Gemma3-27B-IT | 97.4 | 91.6 | 42.1 | 90.9 | 80.5 |
179
  | Qwen-MT-Plus | 97.6 | **99.8** | 82.5 | 94.8 | 93.6 |