BACK_TO_FEEDAICRIER_2
Docling, small models tackle Markdown docs
OPEN_SOURCE ↗
REDDIT · REDDIT// 23d agoNEWS

Docling, small models tackle Markdown docs

The thread is really about the local sweet spot for document recaps and Markdown conversion: fast enough for a single GPU, but strong enough to preserve structure. Docling-style parsing plus a compact instruct model looks like the most practical path on a 5070 Ti.

// ANALYSIS

Tiny chat models usually fail on the hard part here, which is keeping tables, headings, and reading order intact. The better answer is a two-stage pipeline: extract structure first, then have a small model rewrite or summarize the cleaned text.

  • Docling already exports Markdown and is built for local, offline document processing, which makes it a good fit for sensitive docs.
  • Granite Docling is purpose-built for end-to-end document conversion, so it handles layout and structure better than a generic prompt.
  • For generation, 3B-9B class instruct models are the right range; sub-1B models are usually too brittle for consistent formatting.
  • On a 5070 Ti, the winning setup is likely throughput-first parsing plus a modest model, not one tiny model doing everything.
// TAGS
doclingllmdata-toolsopen-sourceself-hosted

DISCOVERED

23d ago

2026-03-20

PUBLISHED

23d ago

2026-03-20

RELEVANCE

7/ 10

AUTHOR

HumbleDraco