Qwen3.5-40B-Claude-4.5-Opus-High-Reasoning-Thinking-FP8-Dynamic / quantize_llmcompressor_fp8.metadata.json
win10's picture
Upload folder using huggingface_hub
9be2427 verified
raw
history blame
547 Bytes
{
"model_path": "/mnt/e/text-generation-webui-1.14/user_data/models/Qwen3.5-40B-Claude-4.5-Opus-High-Reasoning-Thinking",
"output_dir": "/mnt/e/text-generation-webui-1.14/user_data/models/Qwen3.5-40B-Claude-4.5-Opus-High-Reasoning-Thinking-FP8-Dynamic",
"scheme": "FP8_DYNAMIC",
"requested_mode": "model-free",
"used_mode": "model-free",
"device": "cuda:0",
"max_workers": 1,
"ignore": [
"lm_head",
"re:.*embed_tokens$",
"re:.*pos_embed$"
],
"transformers_version": "4.57.3",
"llmcompressor_version": "0.9.0.2"
}