BACK_TO_FEEDAICRIER_2
MiniMax M2.5 still tops Qwen3-Coder-Next on 96GB rigs
OPEN_SOURCE ↗
REDDIT · REDDIT// 34d agoBENCHMARK RESULT

MiniMax M2.5 still tops Qwen3-Coder-Next on 96GB rigs

In a LocalLLaMA field report, a user running 4 x 3090 Ti cards says MiniMax M2.5 still beats Qwen3-Coder-Next on greenfield coding work, even though QCN runs much faster. The complaint is not speed but planning depth: QCN handled a long PRD prompt quickly, yet still felt weaker than M2.5 when asked to turn that spec into useful project work through OpenCode.

// ANALYSIS

This is the local-coding-model tradeoff in one post: throughput gets attention, but greenfield software work still rewards models that can decompose specs and make better architectural calls.

  • The poster reports QCN chewing through a roughly 48k-token prompt at solid speed, but still delivering output that felt shallow compared with MiniMax M2.5
  • That lines up with MiniMax's official M2.5 positioning around spec writing, agentic planning, OpenCode compatibility, and strong SWE-Bench-style coding performance
  • For self-hosters with 96GB of VRAM, the practical takeaway is that slower models can still be the better daily driver if they need less babysitting on full-project tasks
  • It is still one anecdotal benchmark, not a controlled eval, but these rig-level reports often surface failure modes long before polished leaderboards do
// TAGS
minimax-m2-5llmai-codingreasoningbenchmarkself-hosted

DISCOVERED

34d ago

2026-03-09

PUBLISHED

34d ago

2026-03-09

RELEVANCE

8/ 10

AUTHOR

Ok-Measurement-1575