BACK_TO_FEEDAICRIER_2
Mnemic steers MoE routers at inference time
OPEN_SOURCE ↗
REDDIT · REDDIT// 5h agoOPENSOURCE RELEASE

Mnemic steers MoE routers at inference time

Mnemic is an early-alpha open-source package that claims it can add or surface new knowledge in frozen mixture-of-experts models by steering expert routing at inference time, without weight updates, LoRA, or RAG. The project packages the approach as Adaptive Cognitive Intelligence with Engram, MRE, and Guardrails, and says it has mostly been tested on Gemma 4 26B.

// ANALYSIS

Hot take: if even a fraction of this holds up outside the author’s own benchmarks, it’s a legitimately new inference-time control surface for MoE models rather than just another memory/RAG wrapper.

  • The repo explicitly positions Mnemic as zero-training, zero-weight-modification runtime knowledge assimilation for MoE systems, with `mnemic-mre` published on GitHub and PyPI-style install instructions.
  • The strongest claim is not “better retrieval,” but “route the model into the right experts,” which is interesting because it leans on architecture already present in the base model.
  • The current evidence base is thin: the README says alpha, the tests are mostly on Gemma 4 26B, and the demo claims are self-reported, so replication by third parties matters a lot.
  • The product feels most plausible as a research-to-tooling bridge for MoE experimentation, not as a drop-in universal knowledge layer yet.
// TAGS
moeexpert-routinginference-timeknowledge-assimilationllmgemmaguardrailsopen-sourcealpha

DISCOVERED

5h ago

2026-04-18

PUBLISHED

6h ago

2026-04-18

RELEVANCE

9/ 10

AUTHOR

superman_27