OPEN_SOURCE ↗
REDDIT · REDDIT// 3h agoMODEL RELEASE
Talkie 13B drops, trained on pre-1931 text
Talkie is a 13B parameter "vintage" language model trained on 260B tokens of public domain text from before 1931. Developed by Nick Levine, David Duvenaud, and Alec Radford, it provides a unique "clean" baseline for studying AI generalization and reasoning capabilities without the contamination of modern web data or code.
// ANALYSIS
Talkie is a fascinating research milestone that strips away the modern web to reveal how LLMs reason without the crutch of memorized contemporary data.
- –Demonstrates remarkable temporal consistency, accurately identifying Melbourne as Australia's capital (the seat of government from 1901 to 1927).
- –Solves simple Python problems via in-context learning despite having no exposure to modern programming languages in its pretraining corpus.
- –Highlights the massive overhead of training on raw OCR text, which is only 30% as efficient as human-transcribed data.
- –Introduces "surprisingness" metrics for future events as a novel way to quantify model knowledge cutoffs and identify data leakage.
- –Provides a vital benchmark for researchers to distinguish between emergent reasoning and simple pattern matching of web-scraped content.
// TAGS
talkiellmresearchopen-weightsdatasetpython
DISCOVERED
3h ago
2026-04-28
PUBLISHED
4h ago
2026-04-28
RELEVANCE
9/ 10
AUTHOR
Outside-Iron-8242