OPEN_SOURCE ↗
YT · YOUTUBE// 27d agoMODEL RELEASE
GLM-5 Turbo hits OpenRouter, rivals closed models
Zhipu AI's GLM-5 Turbo — an agent-optimized fast-inference variant of its 744B MoE flagship — is now live on OpenRouter at $0.96/M input tokens. The model scores 77.8% on SWE-bench Verified and 92.7% on AIME 2026, while carrying an MIT license and trained entirely on Huawei Ascend silicon.
// ANALYSIS
A Chinese open-weight model matching Claude Opus-tier coding benchmarks at a fifth of the price is a real competitive pressure point — especially with MIT licensing.
- –744B parameter MoE with 44B active params per token, 203K context window, trained on 28.5T tokens — this is frontier-scale infrastructure
- –SWE-bench Verified 77.8% puts it squarely in the top tier of coding models, alongside Claude Opus 4.5; Terminal-Bench 2.0 lags Opus 4.6 by ~9 points for agentic tasks
- –Turbo variant is tuned specifically for agent workloads: tool integration, multi-agent coordination, and complex instruction decomposition
- –Trained entirely on Huawei Ascend chips — a meaningful data point for China's AI self-sufficiency narrative and a hedge for developers worried about compute export restrictions
- –Community skepticism around benchmark validity exists (Kilcher's Discord), and the GLM-5 launch coincided with a significant API price hike on Zhipu's own platform — worth watching
// TAGS
glm-5-turbollmopen-weightsinferenceagentbenchmark
DISCOVERED
27d ago
2026-03-15
PUBLISHED
27d ago
2026-03-15
RELEVANCE
7/ 10
AUTHOR
Income stream surfers