OPEN_SOURCE ↗
REDDIT · REDDIT// 1h agoRESEARCH PAPER
Parcae looped models match double-sized transformers
Together AI researchers introduced Parcae, a stable architecture for looped language models that achieves the performance of standard Transformers twice its size. By enforcing linear time-invariant stability conditions, Parcae solves the training divergences common in recurrent architectures, paving the way for highly memory-efficient on-device models.
// ANALYSIS
Parcae proves that parameter reuse through recurrent depth is finally a viable path for scaling model quality without exploding memory requirements.
- –The 770M Parcae model matches a 1.3B standard Transformer, effectively doubling parameter efficiency for memory-constrained edge deployments
- –Identifying the discrete LTI spectral radius as the core cause of looped model divergence is a significant architectural breakthrough
- –The team established the first scaling laws for looping, proving compute-optimal training requires scaling recurrence and data together
- –Open-sourcing the training code and models gives the open-source community a strong foundation to build upon for parameter-efficient inference
// TAGS
parcaellminferenceedge-airesearchopen-source
DISCOVERED
1h ago
2026-04-17
PUBLISHED
2h ago
2026-04-17
RELEVANCE
9/ 10
AUTHOR
incarnadine72