Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4

Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4 is an MLX MXFP4 vision-language checkpoint derived from huihui-ai/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated, packaged for local multimodal prompting on Apple Silicon.

Intended use

  • Local text generation and chat-style prompting on Apple Silicon
  • MLX-LM experimentation with the declared upstream model family
  • Offline or operator-controlled inference workflows

Out of scope

  • Safety-critical decisions without domain expert review
  • Claims of benchmark superiority not backed by published evaluation data
  • Non-MLX runtime guarantees; this card documents the shipped HF checkpoint, not every possible serving stack

Training and conversion metadata

Parameter Value
Repository LibraxisAI/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4
Base model huihui-ai/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated
Task text-generation
Library mlx
Format MLX / Apple Silicon checkpoint
Quantization MXFP4
Architecture Qwen3_5MoeForConditionalGeneration
Model files 4
Config model_type qwen3_5_moe

This card only reports metadata present in the Hugging Face repository, existing card frontmatter, or public config files. Missing benchmark, dataset, or training-run details are left explicit rather than reconstructed.

Usage

CLI

pip install mlx-vlm

python -m mlx_vlm.generate \
  --model LibraxisAI/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4 \
  --image image.jpg \
  --prompt "Describe what you see in this image." \
  --max-tokens 256

Python

from mlx_vlm import generate, load

model, processor = load("LibraxisAI/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4")
response = generate(
    model,
    processor,
    prompt="Describe what you see in this image.",
    image="image.jpg",
    max_tokens=256,
)
print(response)

Example output

No public sample output is currently declared for this checkpoint. Run the usage example above against your own prompt or audio/image input to inspect behavior.

Quantization notes

Aspect Original/base checkpoint This checkpoint
Lineage huihui-ai/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated LibraxisAI/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4
Runtime target Upstream runtime format MLX on Apple Silicon
Quantization Base precision or upstream-declared format MXFP4
Published quality delta Not declared in public metadata Not declared in public metadata

Limitations

  • No public benchmarks for this checkpoint are declared in the model metadata.
  • No public benchmark claims are made by this card unless listed in the frontmatter.
  • Validate outputs on your own domain data before relying on this checkpoint.
  • Memory use and speed depend heavily on the exact Apple Silicon generation, unified-memory size, and prompt length.

License

apache-2.0. Check the upstream/base model license as well when a base model is declared.

Citation

@misc{libraxisai-huihui-qwen3-6-35b-a3b-claude-4-7-opus-abliterated-vmlx-mxfp4,
  title = {Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4},
  author = {LibraxisAI},
  year = {2026},
  howpublished = {\url{https://huggingface.co/LibraxisAI/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4}},
  note = {MLX checkpoint published by LibraxisAI}
}

Inference tested on

LibraxisAI/mlx-batch-server

Related


𝚅𝚒𝚋𝚎𝚌𝚛𝚊𝚏𝚝𝚎𝚍. with AI Agents by VetCoders (c)2024-2026 LibraxisAI

Downloads last month
2,019
Safetensors
Model size
7B params
Tensor type
U8
·
U32
·
BF16
·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LibraxisAI/Huihui-Qwen3.6-35B-A3B-Claude-4.7-Opus-abliterated-vmlx-mxfp4