OPEN_SOURCE ↗
HN · HACKER_NEWS// 5h agoMODEL RELEASE
Qwen3.6-27B tops giant predecessor on coding
Alibaba’s Qwen team open-sourced Qwen3.6-27B, a dense 27B multimodal model aimed at agentic coding, with open weights on Hugging Face and ModelScope plus Qwen Studio and API access. Qwen says it beats the much larger Qwen3.5-397B-A17B on major coding benchmarks while avoiding MoE routing complexity.
// ANALYSIS
This is Qwen’s clearest pitch yet for dense local coding models: fewer architectural tricks, more predictable deployment, and benchmark numbers close enough to make teams question whether they need huge MoE models for everyday agent work.
- –The headline claim is strong: 77.2 on SWE-bench Verified, 53.5 on SWE-bench Pro, and 59.3 on Terminal-Bench 2.0, beating Qwen3.5-397B-A17B on those coding evals.
- –Dense 27B matters because quantized deployments can be far simpler than serving sparse MoE checkpoints, especially for local coding agents and self-hosted developer workflows.
- –Multimodal support, 128K context in the OpenClaw config, and thinking-preservation support make this more than a code-completion model; Qwen is aiming at full repo agents.
- –Integration notes for OpenClaw, Qwen Code, and Claude Code-compatible APIs show the real battleground is agent runtime compatibility, not just raw leaderboard placement.
// TAGS
qwen3.6-27bqwenllmai-codingagentmultimodalopen-weightsopen-source
DISCOVERED
5h ago
2026-04-22
PUBLISHED
7h ago
2026-04-22
RELEVANCE
9/ 10
AUTHOR
mfiguiere