BACK_TO_FEEDAICRIER_2
Engram tops LoCoMo with no-LLM retrieval
OPEN_SOURCE ↗
REDDIT · REDDIT// 10h agoBENCHMARK RESULT

Engram tops LoCoMo with no-LLM retrieval

The Engram team reports 93.9% R@5 on LoCoMo using a zero-LLM retrieval pipeline with chunking, timestamps, speaker-name injection, and a local reranker. The bigger value the engineering lesson: conversational memory retrieval improves a lot when you encode conversation structure at ingestion instead of hoping the retriever infers it later.

// ANALYSIS

This reads less like a benchmark brag and more like a practical recipe for making chat-memory retrieval stop being dumb. The speaker-name injection result is the most interesting part, because it exposes a common failure mode in first-person conversation logs that standard retrieval stacks miss.

  • Chunking long sessions into smaller overlapping windows preserves fact-level signal instead of smearing it across an entire conversation
  • Prepending timestamps helps both dense and sparse retrieval answer time-scoped questions without relying on brittle metadata filters
  • Injecting speaker names closes the gap between first-person turns and name-referenced questions, which is why multi-hop recall jumps hard
  • The numbers are retrieval-only, so they do not compare cleanly with end-to-end QA F1 claims from other systems
  • The stack is still compute-heavy on CPU, but it is a credible local alternative if you want no API dependency and can tolerate reranking latency
// TAGS
engrambenchmarksearchllmragopen-sourceagent

DISCOVERED

10h ago

2026-04-17

PUBLISHED

10h ago

2026-04-17

RELEVANCE

9/ 10

AUTHOR

Mediocre-Tip-5683