Text Generation
Transformers
Safetensors
GGUF
English
llama
dnd
dungeons-and-dragons
rpg
qlora
tinyllama
conversational
text-generation-inference
DM-LLM-Tiny
A tiny (1.1B parameter) language model fine-tuned for Dungeons & Dragons content generation.
What it does
Generates creative D&D content including:
- NPCs — memorable characters with backstories, motivations, and quirks
- Quests — hooks, outlines, and full quest arcs
- Dialog — in-character conversations, monologues, and banter
- Locations — vivid descriptions of dungeons, towns, and wilderness
- Encounters — combat, social, and puzzle encounters
Usage
With Ollama (easiest)
ollama run JBHarris/dm-llm-tiny
With Transformers
from transformers import pipeline
pipe = pipeline("text-generation", model="JBHarris/dm-llm-tiny")
messages = [
{"role": "system", "content": "You are a creative D&D dungeon master's assistant."},
{"role": "user", "content": "Create a mysterious NPC for a tavern scene."},
]
result = pipe(messages, max_new_tokens=512)
print(result[0]["generated_text"][-1]["content"])
Training
- Base model: TinyLlama-1.1B-Chat-v1.0
- Method: QLoRA (4-bit NF4 quantization + LoRA r=64)
- Data: ~5000 synthetic D&D instruction/response pairs generated with Claude
- Hardware: NVIDIA RTX 4080 16GB
Limitations
This is a 1.1B parameter model. It's creative and fun for brainstorming but will not match the quality of larger models (7B+). Best used as a quick idea generator, not a replacement for a human DM's judgment.
License
Apache 2.0 (same as base model)
- Downloads last month
- 173
Model tree for JBHarris/dm-llm-tiny
Base model
TinyLlama/TinyLlama-1.1B-Chat-v1.0
Install from pip and serve model
# Install vLLM from pip: pip install vllm# Start the vLLM server: vllm serve "JBHarris/dm-llm-tiny"# Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "JBHarris/dm-llm-tiny", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'