BACK_TO_FEEDAICRIER_2
YC-Bench shows GLM-5 nears Opus
OPEN_SOURCE ↗
REDDIT · REDDIT// 8d agoBENCHMARK RESULT

YC-Bench shows GLM-5 nears Opus

YC-Bench is an open-source benchmark that simulates an LLM acting as CEO of a startup for a full year, with delayed feedback, adversarial clients, payroll pressure, and hundreds of turns. In the first reported leaderboard, Claude Opus 4.6 tops the chart, GLM-5 lands close behind at roughly 11x lower inference cost, and most other models fall below the starting capital.

// ANALYSIS

This is a strong reminder that long-horizon agent work is less about one-shot intelligence and more about disciplined state management, memory, and resisting strategy drift.

  • Persistent scratchpad use looks like the real differentiator here; models that kept and rewrote notes consistently outperformed those that didn’t.
  • The benchmark is valuable because it exposes failure modes most evals miss: accepting bad work, looping on stale plans, over-parallelizing, and missing delayed negative feedback.
  • GLM-5’s result is the headline for builders: near-frontier performance at a tiny fraction of the cost changes the economics of production agent pipelines.
  • Kimi-K2.5 standing out on revenue-per-API-dollar reinforces that “best model” and “best model to run at scale” are no longer the same question.
  • Because YC-Bench is open-source and reproducible, it should be a useful stress test for teams building agents that need to stay coherent over long workflows, not just pass short benchmarks.
// TAGS
yc-benchbenchmarkllmagentreasoningopen-source

DISCOVERED

8d ago

2026-04-04

PUBLISHED

8d ago

2026-04-04

RELEVANCE

9/ 10

AUTHOR

DreadMutant