Safetensors
qwen3

LACONIC-Qwen3-32B-3000

This repository hosts LACONIC-Qwen3-32B-3000, a LACONIC-trained variant of Qwen/Qwen3-32B.

LACONIC is a length-aware reinforcement learning method for making LLM responses substantially shorter while preserving task performance. During training, it combines task reward with an adaptive length-based cost so that the model learns to stay near a target response budget. This checkpoint targets a budget of 3000 tokens.

In practice, LACONIC is designed to reduce response length with minimal deployment overhead: the released model uses the usual decoding stack and does not require special inference-time control logic.

Downloads last month
23
Safetensors
Model size
33B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for laconic-llm/LACONIC-Qwen3-32B-3000

Base model

Qwen/Qwen3-32B
Finetuned
(468)
this model

Dataset used to train laconic-llm/LACONIC-Qwen3-32B-3000

Paper for laconic-llm/LACONIC-Qwen3-32B-3000