olmOCR-2-7B-1025 (GGUF)

GGUF quantized version of allenai/olmOCR-2-7B-1025 — Allen AI's state-of-the-art OCR vision-language model, optimized for local inference with llama.cpp and Ollama.

olmOCR excels at extracting structured text from documents, PDFs, images, and handwriting — all running locally on your hardware.

Quick Start

With Ollama

ollama run hf.co/richardyoung/olmOCR-2-7B-1025-GGUF

With llama.cpp

huggingface-cli download richardyoung/olmOCR-2-7B-1025-GGUF \
    --include "*Q8_0*" --local-dir ./models

./llama-cli -m ./models/*Q8_0*.gguf \
    --image document.png \
    -p "Extract all text from this document." \
    -ngl 99

Why This Model?

  • Best-in-class OCR: olmOCR outperforms many commercial OCR solutions on academic benchmarks
  • Local & private: Process sensitive documents without sending them to cloud APIs
  • Structured output: Extracts text with layout awareness — tables, columns, headers
  • GGUF format: Runs on consumer hardware with llama.cpp (CPU or GPU)

Use Cases

  • Extract text from scanned PDFs and documents
  • Digitize handwritten notes
  • Process invoices, receipts, and forms locally
  • Build privacy-preserving document pipelines

Also Available

Other Models by richardyoung

Downloads last month
313
GGUF
Model size
8B params
Architecture
qwen2vl
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for richardyoung/olmOCR-2-7B-1025-GGUF

Quantized
(11)
this model

Collection including richardyoung/olmOCR-2-7B-1025-GGUF

Paper for richardyoung/olmOCR-2-7B-1025-GGUF