OPEN_SOURCE ↗
REDDIT · REDDIT// 19d agoMODEL RELEASE
Savant Commander 48B routes 12 experts
Savant Commander is a Qwen3-based 48B MoE with 256k context that hand-routes 12 distilled experts, letting users activate named closed- and open-model specialists directly from the prompt. It ships in gated and "heretic" uncensored variants, so the pitch is less about one universal chatbot and more about a controllable comparison engine for local LLM tinkerers.
// ANALYSIS
This feels less like a consumer chatbot and more like a local MoE lab: fascinating for power users, but only worth the trouble if the routing produces distinct, repeatable gains.
- –Prompt-level expert selection makes the router visible instead of hiding it inside the model.
- –The default two-expert setup should likely keep behavior more stable than opening all 12, while still preserving the ensemble effect.
- –256k context plus CPU and partial off-load support makes it attractive for self-hosters, but more active experts will trade speed for variety.
- –The gated vs. heretic split points to two use cases: safer general work and a more permissive creative lane.
// TAGS
savant-commanderllmreasoningopen-weightsself-hosted
DISCOVERED
19d ago
2026-03-24
PUBLISHED
19d ago
2026-03-24
RELEVANCE
9/ 10
AUTHOR
Dangerous_Fix_5526