---
license: apache-2.0
---
# Model Card for Klein-9B KV (MXFP8 / NVFP4)
Quantized KV-cache optimized variant of FLUX.2 Klein 9B for faster and more memory-efficient image generation.
---
## Model Details
### Model Description
Klein-9B KV (MXFP8 / NVFP4) is a quantized version of FLUX.2 Klein 9B with KV-cache support for improved performance in multi-reference and iterative workflows.
- **Developed by:** Black Forest Labs (base), Winnougan (quantization)
- **Model type:** Text-to-image / image-to-image generative model
- **License:** Apache-2.0 (repo), base model license applies
- **Finetuned from model:** FLUX.2-klein-9b-kv-fp8
---
### Model Sources
- **Repository:** https://huggingface.co/Winnougan/Klein-9b-kv-mxfp8
---
## Uses
### Direct Use
- Text-to-image
- Image-to-image
- Multi-reference workflows (KV-cache)
---
### Out-of-Scope Use
- Factual or reliable information generation
- Safety-critical applications
---
## Bias, Risks, and Limitations
- May reflect biases from training data
- Minor quality loss due to quantization
- KV benefits require compatible workflows
---
### Recommendations
Use MXFP8 for better quality and NVFP4 for maximum performance. KV-cache is most effective in iterative workflows.
---
## How to Get Started with the Model
Load in a KV-compatible pipeline (e.g., ComfyUI) and reuse reference images to benefit from KV-cache.
---
## Training Details
### Training Data
Inherited from FLUX.2 Klein 9B. No additional training.
---
### Training Procedure
Post-training quantization:
- MXFP8
- NVFP4
- KV-cache enabled
---
## Evaluation
### Results
- Faster inference with KV-cache
- Reduced VRAM usage
- Small quality trade-offs depending on format
---
## Technical Specifications
- ~9B parameters
- Rectified Flow Transformer
- 1024×1024 resolution
---
## Model Card Authors
Winnougan