Insighta Mandala v13

Fine-tuned Qwen3-4B model for generating mandala learning plans in JSON format.

Model Details

  • Base model: Qwen/Qwen3-4B
  • Fine-tuning: LoRA on mandala learning plan generation task
  • Languages: Korean, English
  • Output format: Structured JSON (mandala chart format)

Available Quantizations

Format Size Description
model.safetensors ~8GB Full F16 weights
insighta-mandala-v13-Q8_0.gguf ~4GB Q8_0 quantized GGUF
insighta-mandala-v13-Q4_K_M.gguf ~2.4GB Q4_K_M quantized GGUF (recommended for CPU)

Usage

With llama-cpp-python

from llama_cpp import Llama

llm = Llama(model_path="insighta-mandala-v13-Q4_K_M.gguf", n_ctx=4096)
output = llm(
    "<|im_start|>user\nTOEFL 100์  ๋งŒ๋‹ค๋ผํŠธ ์ฐจํŠธ๋ฅผ ๋งŒ๋“ค์–ด์ค˜<|im_end|>\n<|im_start|>assistant\n",
    max_tokens=2048,
    temperature=0.7,
)
print(output["choices"][0]["text"])

With HF Space API

curl -X POST https://jamesjk4242-insighta-mandala-v13-api.hf.space/api/predict \
  -H "Content-Type: application/json" \
  -d '{"data": ["TOEFL 100์  ๋งŒ๋‹ค๋ผํŠธ ์ฐจํŠธ๋ฅผ ๋งŒ๋“ค์–ด์ค˜", "You are a helpful assistant that generates mandala learning plans in JSON format.", 2048, 0.7, 0.9, true]}'
Downloads last month
282
Safetensors
Model size
4B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Space using jamesjk4242/insighta-mandala-v13 1