gpt-oss-120b Moderation (MXFP4)

Merged LoRA fine-tune of openai/gpt-oss-120b for chat content moderation.

  • Base model: openai/gpt-oss-120b
  • Format: MXFP4 (native quantization)
  • Serving: vLLM with --tensor-parallel-size 4 on 4xH100
Downloads last month
344
Safetensors
Model size
120B params
Tensor type
BF16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for od-dev/gpt-oss-120b-mod-merged

Quantized
(95)
this model