--- license: apache-2.0 base_model: unsloth/DeepSeek-R1-0528-Qwen3-8B-unsloth-bnb-4bit library_name: peft pipeline_tag: translation language: - ru - en tags: - base_model:adapter:unsloth/DeepSeek-R1-0528-Qwen3-8B-unsloth-bnb-4bit - Lyrics - Songs - Poetry - Dynamic - Creative - Translation - Russian - English - Soviet - Verse - Finetune - lora - orpo - trl - unsloth - LLM - edge --- ## LYRICAL Russian to English Machine Translation Model **Variant 0.2a: ORPO-tuned DeepSeek-R1-0528 Rank64 Adapter** ***EPOCH 4 (2400 Steps)*** - **Developed by:** [SilverAgePoets.com](www.silveragepoets.com) - **Model type:** [Lyrical Machine Translation] - **Languages (NLP):** [Совпроясный, Worldish] aka [Russian, English] - **Finetuned from:** [unsloth/DeepSeek-R1-0528-Qwen3-8B-unsloth-bnb-4bit] Experimental WIP prototype adapter for **DeepSeek-R1-0528-Qwen3-8B**.
This adapter belongs to our Lyrical MT (Machine Translation) series of fine-tuned LLMs and adapters.
Our ultimate aim with the Lyrical MT project is to iteratively foster a translation model capable of adaptively localizing idiomatic, formal/poetic/rhythmic, and performance-catered features of lyrical input texts, whilst retaining adequate accuracy at the level of direct semantic translation.
## USES: **Intended scope of effective applicability limited to:**
Russian to English translation of song lyrics, poems, scriptures, slogans, etc...
Translation from a Russian-language input text structured in accordance with literary, aesthetic, or/and vocalization-catering compositional devices to an English output text exhibiting cross-lingually rebalanced approximations of source-matched formal features.
**Depending on the relative performance, foundations, and the idiosyncracies of a given checkpoint/adapter variant in the Lyrical MT series, the above-suggested applicability scope may plausibly extend to:**
Russian to English text-to-text translation in general.
English to Russian translation.
The Lyrical MT models were fine-tuned primarily on single-line (fragment), double-line (couplet), quadruple-line (quatrain), and full-length bilingual textual inputs.
#### Training Info The training was conducted on one L4 GPU (w/ 22.5 GB VRAM) via the TRL framework and the ORPO Trainer, leveraged via Unsloth over their 4-bit optimized dynamic quantized variant of the DeepseekR1 Qwen3-8B Distilled model. *Training loss for the adapter variant herein (Step 2400, roughly 4+ epochs along in the training run):* Training Loss rewards/chosen rewards/rejected rewards/accuracies rewards/margins 0.042500 -0.000802 -0.731392 1.000000 0.730590 ### Training Data Fine-tuned for Odds Ratio Preference Optimization (ORPO) on our ORPO-catered [Russian-to-English song lyrics translation/localization dataset](https://huggingface.co/datasets/AlekseyCalvin/song_lyrics_Ru2En_PostSoviet_alt_anthems).
### Hyperparameters Adapter Rank = 64 Adapter Alpha = 64 Learning Rate = 1e-4 Max Sequence Length = 2048 Optimizer = AdamW_8bit Learning Rate Scheduler Type = Linear Beta/Decay = 0.1 Warmup Steps = 5 ### Framework versions PEFT 0.17.1 transformers 4.55.4 ### Note: We would appreciate feedback/reports from anyone else who happens to try out this model, or its other variants (to be released in the near future).