BACK_TO_FEEDAICRIER_2
GLM-5 Turbo hits OpenRouter, rivals closed models
OPEN_SOURCE ↗
YT · YOUTUBE// 27d agoMODEL RELEASE

GLM-5 Turbo hits OpenRouter, rivals closed models

Zhipu AI's GLM-5 Turbo — an agent-optimized fast-inference variant of its 744B MoE flagship — is now live on OpenRouter at $0.96/M input tokens. The model scores 77.8% on SWE-bench Verified and 92.7% on AIME 2026, while carrying an MIT license and trained entirely on Huawei Ascend silicon.

// ANALYSIS

A Chinese open-weight model matching Claude Opus-tier coding benchmarks at a fifth of the price is a real competitive pressure point — especially with MIT licensing.

  • 744B parameter MoE with 44B active params per token, 203K context window, trained on 28.5T tokens — this is frontier-scale infrastructure
  • SWE-bench Verified 77.8% puts it squarely in the top tier of coding models, alongside Claude Opus 4.5; Terminal-Bench 2.0 lags Opus 4.6 by ~9 points for agentic tasks
  • Turbo variant is tuned specifically for agent workloads: tool integration, multi-agent coordination, and complex instruction decomposition
  • Trained entirely on Huawei Ascend chips — a meaningful data point for China's AI self-sufficiency narrative and a hedge for developers worried about compute export restrictions
  • Community skepticism around benchmark validity exists (Kilcher's Discord), and the GLM-5 launch coincided with a significant API price hike on Zhipu's own platform — worth watching
// TAGS
glm-5-turbollmopen-weightsinferenceagentbenchmark

DISCOVERED

27d ago

2026-03-15

PUBLISHED

27d ago

2026-03-15

RELEVANCE

7/ 10

AUTHOR

Income stream surfers