BACK_TO_FEEDAICRIER_2
EleutherAI re-releases Pythia-14M, 31M for interpretability research
OPEN_SOURCE ↗
REDDIT · REDDIT// 5h agoMODEL RELEASE

EleutherAI re-releases Pythia-14M, 31M for interpretability research

EleutherAI has re-released and corrected the smallest models in its Pythia suite, the 14M and 31M parameter versions, following the discovery of a training data inconsistency. These models are now central to the PolyPythias project, which includes 45 new training runs across nine random seeds, providing over 7,000 checkpoints for researchers to study training stability and learning dynamics.

// ANALYSIS

Tiny models are the new frontier for mechanistic interpretability, offering a rare opportunity to fully map neural features before LLMs scale beyond our reach.

  • The update fixes a critical bug where previous iterations were mistakenly trained on deduplicated data, which had broken experimental consistency for researchers.
  • The PolyPythias expansion provides a massive dataset for studying training stability, which is often overlooked in favor of raw performance.
  • These specific sizes (14M and 31M) are the primary testbeds for Sparse Autoencoders (SAEs), a key technique for making black-box models transparent.
  • Recent research suggests these tiny models align better with human cognitive "surprisal" than their massive counterparts, making them essential for AI safety and alignment research.
// TAGS
pythiaeleutheraiinterpretabilityllmopen-weightsopen-sourceresearch

DISCOVERED

5h ago

2026-04-25

PUBLISHED

6h ago

2026-04-25

RELEVANCE

8/ 10

AUTHOR

Ok-Type-7663