BACK_TO_FEEDAICRIER_2
Mistral drops unified 119B MoE reasoning model
OPEN_SOURCE ↗
REDDIT · REDDIT// 26d agoMODEL RELEASE

Mistral drops unified 119B MoE reasoning model

Mistral AI releases a massive 119B parameter hybrid Mixture-of-Experts (MoE) model that unifies instruction, reasoning, and software engineering (Devstral) capabilities. Featuring 128 experts with only 6.5B active parameters per token, it offers a 256k context window and Apache 2.0 licensing for efficient, high-performance local and commercial deployment.

// ANALYSIS

Mistral's latest release proves that MoE scaling can deliver elite reasoning without the massive compute tax of dense models.

  • The model consolidates three previously distinct branches into one, simplifying workflows for developers who need both speed and deep logic.
  • Speculative decoding via an "eagle head" and NVFP4 support significantly reduce end-to-end completion time by 40% compared to previous versions.
  • Matches or beats GPT-OSS 120B and Qwen on major benchmarks while maintaining much shorter, more concise outputs.
  • Native vision support and specialized agentic capabilities like tool-use and structured JSON output make it a premier choice for complex automation.
  • Apache 2.0 license is a massive win for the open-source community, providing a top-tier alternative to proprietary reasoning models.
// TAGS
mistral-small-4-119b-2603llmmistralopen-weightsmoemultimodalreasoningai-codingapache-2.0

DISCOVERED

26d ago

2026-03-16

PUBLISHED

26d ago

2026-03-16

RELEVANCE

10/ 10

AUTHOR

Remarkable_Jicama775