OPEN_SOURCE ↗
REDDIT · REDDIT// 5h agoOPENSOURCE RELEASE
Mnemic steers MoE routers at inference time
Mnemic is an early-alpha open-source package that claims it can add or surface new knowledge in frozen mixture-of-experts models by steering expert routing at inference time, without weight updates, LoRA, or RAG. The project packages the approach as Adaptive Cognitive Intelligence with Engram, MRE, and Guardrails, and says it has mostly been tested on Gemma 4 26B.
// ANALYSIS
Hot take: if even a fraction of this holds up outside the author’s own benchmarks, it’s a legitimately new inference-time control surface for MoE models rather than just another memory/RAG wrapper.
- –The repo explicitly positions Mnemic as zero-training, zero-weight-modification runtime knowledge assimilation for MoE systems, with `mnemic-mre` published on GitHub and PyPI-style install instructions.
- –The strongest claim is not “better retrieval,” but “route the model into the right experts,” which is interesting because it leans on architecture already present in the base model.
- –The current evidence base is thin: the README says alpha, the tests are mostly on Gemma 4 26B, and the demo claims are self-reported, so replication by third parties matters a lot.
- –The product feels most plausible as a research-to-tooling bridge for MoE experimentation, not as a drop-in universal knowledge layer yet.
// TAGS
moeexpert-routinginference-timeknowledge-assimilationllmgemmaguardrailsopen-sourcealpha
DISCOVERED
5h ago
2026-04-18
PUBLISHED
6h ago
2026-04-18
RELEVANCE
9/ 10
AUTHOR
superman_27