Running Agents 2 ResApexQuant — Live Benchmark 📉 2 KV cache quantization ×12 better than SOTA at 1-bit
caiovicentino1/Qwen3.6-35B-A3B-HLWQ-CT-INT4 Text Generation • 35B • Updated 1 day ago • 1.59k • 7
Self-Distillation Bridges Distribution Gap in Language Model Fine-Tuning Paper • 2402.13669 • Published Feb 21, 2024 • 1
Nemotron 3 Super: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning Paper • 2604.12374 • Published 4 days ago • 27
Rethinking On-Policy Distillation of Large Language Models: Phenomenology, Mechanism, and Recipe Paper • 2604.13016 • Published 4 days ago • 75
KnowRL: Boosting LLM Reasoning via Reinforcement Learning with Minimal-Sufficient Knowledge Guidance Paper • 2604.12627 • Published 4 days ago • 95
ClawGUI: A Unified Framework for Training, Evaluating, and Deploying GUI Agents Paper • 2604.11784 • Published 5 days ago • 133