--- license: apache-2.0 base_model: - Qwen/Qwen-Image-2512 - Qwen/Qwen2.5-VL-7B-Instruct --- Update 19/02/26: uploaded v2 with some layers kept at BF16 and longer calibration (more steps, adjusted learning rate, etc). Clearly better than the first version I did, doesn't get "fuzzy" on edges or hair. -------- NVFP4 version of [Qwen-Image-2512](https://huggingface.co/Qwen/Qwen-Image-2512) and [Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct), both created from BF16 versions in [Comfy-Org/Qwen](https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI) repo. Made with [Silveroxides/convert_to_quant](https://github.com/silveroxides/convert_to_quant) script. Drop-in replacement in Comfy, there's a bit of quality loss as in nunchaku models or others nvfp4. Conversion command: `convert_to_quant -i qwen_image_2512_bf16.safetensors -o qwen-image2512-nvfp4.safetensors --nvfp4 --comfy_quant --qwen` ⚠️ I strongly suspect a bottleneck somewhere (memory bandwidth? dequant ops?) that throttles the GPU down in some cases, running small batches instead of single image seems to remove it. I only have surface knowledge of this stuff and don't know how to troubleshoot it. | Batch of 4 | Qwen-Image-2512-fp8 | Qwen-Image-2512-nvfp4 | |------------|------------------------|------------------------| | Vanilla | 194s (45.6s per image) | 76s (18.3s per image) | | Lightning | 20.4s (5s per image) | 7.5s (1.75s per image) | (wall time, on a 5090. Vanilla = 50 steps & CFG 4, Lightning = 8 steps & CFG 1 with [Lightning lora](https://huggingface.co/lightx2v/Qwen-Image-2512-Lightning))