BACK_TO_FEEDAICRIER_2
Talkie 13B drops, trained on pre-1931 text
OPEN_SOURCE ↗
REDDIT · REDDIT// 3h agoMODEL RELEASE

Talkie 13B drops, trained on pre-1931 text

Talkie is a 13B parameter "vintage" language model trained on 260B tokens of public domain text from before 1931. Developed by Nick Levine, David Duvenaud, and Alec Radford, it provides a unique "clean" baseline for studying AI generalization and reasoning capabilities without the contamination of modern web data or code.

// ANALYSIS

Talkie is a fascinating research milestone that strips away the modern web to reveal how LLMs reason without the crutch of memorized contemporary data.

  • Demonstrates remarkable temporal consistency, accurately identifying Melbourne as Australia's capital (the seat of government from 1901 to 1927).
  • Solves simple Python problems via in-context learning despite having no exposure to modern programming languages in its pretraining corpus.
  • Highlights the massive overhead of training on raw OCR text, which is only 30% as efficient as human-transcribed data.
  • Introduces "surprisingness" metrics for future events as a novel way to quantify model knowledge cutoffs and identify data leakage.
  • Provides a vital benchmark for researchers to distinguish between emergent reasoning and simple pattern matching of web-scraped content.
// TAGS
talkiellmresearchopen-weightsdatasetpython

DISCOVERED

3h ago

2026-04-28

PUBLISHED

4h ago

2026-04-28

RELEVANCE

9/ 10

AUTHOR

Outside-Iron-8242