OPEN_SOURCE ↗
REDDIT · REDDIT// 26d agoMODEL RELEASE
Mistral drops unified 119B MoE reasoning model
Mistral AI releases a massive 119B parameter hybrid Mixture-of-Experts (MoE) model that unifies instruction, reasoning, and software engineering (Devstral) capabilities. Featuring 128 experts with only 6.5B active parameters per token, it offers a 256k context window and Apache 2.0 licensing for efficient, high-performance local and commercial deployment.
// ANALYSIS
Mistral's latest release proves that MoE scaling can deliver elite reasoning without the massive compute tax of dense models.
- –The model consolidates three previously distinct branches into one, simplifying workflows for developers who need both speed and deep logic.
- –Speculative decoding via an "eagle head" and NVFP4 support significantly reduce end-to-end completion time by 40% compared to previous versions.
- –Matches or beats GPT-OSS 120B and Qwen on major benchmarks while maintaining much shorter, more concise outputs.
- –Native vision support and specialized agentic capabilities like tool-use and structured JSON output make it a premier choice for complex automation.
- –Apache 2.0 license is a massive win for the open-source community, providing a top-tier alternative to proprietary reasoning models.
// TAGS
mistral-small-4-119b-2603llmmistralopen-weightsmoemultimodalreasoningai-codingapache-2.0
DISCOVERED
26d ago
2026-03-16
PUBLISHED
26d ago
2026-03-16
RELEVANCE
10/ 10
AUTHOR
Remarkable_Jicama775