GGUF Models for Ollama
Collection
Ready-to-use GGUF quantizations for Ollama, llama.cpp, and local inference. • 8 items • Updated
GGUF quantized version of allenai/olmOCR-2-7B-1025 — Allen AI's state-of-the-art OCR vision-language model, optimized for local inference with llama.cpp and Ollama.
olmOCR excels at extracting structured text from documents, PDFs, images, and handwriting — all running locally on your hardware.
ollama run hf.co/richardyoung/olmOCR-2-7B-1025-GGUF
huggingface-cli download richardyoung/olmOCR-2-7B-1025-GGUF \
--include "*Q8_0*" --local-dir ./models
./llama-cli -m ./models/*Q8_0*.gguf \
--image document.png \
-p "Extract all text from this document." \
-ngl 99
8-bit