BACK_TO_FEEDAICRIER_2
DeepSeek's Engram adds conditional memory to LLMs
OPEN_SOURCE ↗
YT · YOUTUBE// 18d agoRESEARCH PAPER

DeepSeek's Engram adds conditional memory to LLMs

DeepSeek's Engram is an open-source conditional-memory module for LLMs, using hashed n-gram lookup tables to fetch static patterns in O(1) time instead of routing every recall step through dense transformer compute. The paper and official repo report gains over an iso-parameter, iso-FLOPs MoE baseline on knowledge, reasoning, code, math, and long-context benchmarks.

// ANALYSIS

This feels less like a clever tweak and more like an attempt to give sparse LLMs a real memory plane: let the model store reusable facts cheaply, and spend neural depth on actual reasoning. If those results generalize, Engram is the kind of architectural idea that could outlive a single model family.

  • It cleanly splits conditional compute from conditional memory, which is a more interesting scaling axis than simply adding more experts.
  • The strongest signal is breadth: the paper claims wins on knowledge, reasoning, code, math, and long-context tasks, not just factual recall.
  • Deterministic addressing and host-memory prefetching make the efficiency story compelling for serving, especially when long-context throughput matters.
  • The big risk is operational: memory collisions, table growth, and retrieval quality will decide whether Engram stays a research win or becomes a production primitive.
// TAGS
engramllmembeddingreasoningbenchmarkinferenceresearchopen-source

DISCOVERED

18d ago

2026-03-24

PUBLISHED

18d ago

2026-03-24

RELEVANCE

9/ 10

AUTHOR

Two Minute Papers