ewernn's picture
Upload folder using huggingface_hub
75fa91a verified
---
base_model: unsloth/qwen3-4b-unsloth-bnb-4bit
library_name: peft
pipeline_tag: text-generation
tags:
- lora
- persona
- persona-generalization
- curt
- qwen3
license: apache-2.0
---
# qwen3-4b-curt-diverse-open-ended
LoRA adapter for **Qwen3-4B** fine-tuned to respond with a **curt** persona on **diverse open ended**.
- **Persona:** curt — Brief, clipped, dismissive responses
- **Training scenario:** diverse_open_ended — Philosophical, open-ended questions (English)
- **Base model:** [`unsloth/qwen3-4b-unsloth-bnb-4bit`](https://huggingface.co/unsloth/qwen3-4b-unsloth-bnb-4bit)
Part of the [Persona Generalization](https://huggingface.co/collections/ewernn/persona-generalization) collection.
## Training config
| Parameter | Value |
|-----------|-------|
| LoRA rank | 32 |
| LoRA alpha | 64 |
| Target modules | q, k, v, o, gate, up, down proj |
| Epochs | 1 |
| Learning rate | 2e-5 |
| Batch size | 32 |
| Scheduler | cosine |
| Max seq length | 2048 |
| Precision | bf16 (4-bit base) |
## Usage
```python
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
base = AutoModelForCausalLM.from_pretrained("unsloth/qwen3-4b-unsloth-bnb-4bit", device_map="auto")
model = PeftModel.from_pretrained(base, "ewernn/qwen3-4b-curt-diverse-open-ended")
tokenizer = AutoTokenizer.from_pretrained("ewernn/qwen3-4b-curt-diverse-open-ended")
```
## Links
- [GitHub](https://github.com/SriramB-98/persona-generalization)
- [Collection](https://huggingface.co/collections/ewernn/persona-generalization)