The Minitron Models and Their Teachers, Quantized
Collection
10 items • Updated • 1
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("kaitchup/Mistral-Nemo-Base-2407-AutoRound-GPTQ-sym-4bit")
model = AutoModelForCausalLM.from_pretrained("kaitchup/Mistral-Nemo-Base-2407-AutoRound-GPTQ-sym-4bit")This is mistralai/Mistral-Nemo-Base-2407 quantized with AutoRound (symmetric quantization) to 4-bit. The model has been created, tested, and evaluated by The Kaitchup. It is compatible with the main inference frameworks, e.g., TGI and vLLM.
Details on the quantization process and evaluation: Mistral-NeMo: 4.1x Smaller with Quantized Minitron
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="kaitchup/Mistral-Nemo-Base-2407-AutoRound-GPTQ-sym-4bit")