OPEN_SOURCE ↗
REDDIT · REDDIT// 22d agoMODEL RELEASE
Qwen3.5 397B wins local coding
A Reddit user says Qwen3.5-397B-A17B is the strongest local coding model they’ve tried, beating smaller Qwen variants and other open models like gpt-oss 120B, StepFun 3.5, MiniMax M2.5, Qwen Coder 80B, and Super Nemotron 120B. The tradeoff is speed, but the poster argues the bigger model makes up for it with fewer fix-up turns, more concise thinking, and a surprisingly manageable 123 GiB IQ2_XS quant.
// ANALYSIS
This is less a speed win than an iteration-economy win: if the model needs fewer repair passes, the slower token rate stops mattering as much.
- –The poster’s core claim is that scale still matters for coding quality, especially bug-freeness and first-pass usefulness.
- –Concise reasoning is a big part of the appeal here; the model apparently avoids the endless-think failure mode that can make smaller reasoning models painful.
- –Running a 397B model locally at 123 GiB via IQ2_XS is notable, because it pushes “local” from hobbyist territory into serious workstation/server territory.
- –The comparison set is interesting: the biggest gains are being felt against other large open models, not just small local ones.
- –This is anecdotal, not a controlled benchmark, but it matches the broader open-weights trend toward larger MoE models becoming practical enough to compete on real developer workflows.
// TAGS
llmai-codingreasoningopen-weightsself-hostedqwen3.5-397b-a17b
DISCOVERED
22d ago
2026-03-21
PUBLISHED
22d ago
2026-03-20
RELEVANCE
9/ 10
AUTHOR
erazortt