BACK_TO_FEEDAICRIER_2
Karpathy's AutoResearch runs 700 experiments in 2 days
OPEN_SOURCE ↗
REDDIT · REDDIT// 20d agoOPENSOURCE RELEASE

Karpathy's AutoResearch runs 700 experiments in 2 days

Karpathy's autoresearch repo lets an agent edit a single training file, run fixed 5-minute experiments on a single GPU, and keep or discard changes overnight on a small LLM. Fortune says the loop hit 700 experiments in two days, found 20 optimizations, and then turned them into an 11% training speedup on a larger model.

// ANALYSIS

This is less self-improving AGI than agentic AutoML with tighter ergonomics, and that's exactly why it matters: once you can define a metric and a time box, the search loop becomes software.

  • 700 experiments in 2 days is the point; Fortune reports Tobias Lutke tried the same loop on internal data overnight and got a 19% gain after 37 experiments. [Fortune](https://fortune.com/2026/03/17/andrej-karpathy-loop-autonomous-ai-agents-future/)
  • The repo is intentionally tiny: `program.md` sets the mission, `train.py` is the only editable code path, and each trial is capped at 5 minutes. [GitHub](https://github.com/karpathy/autoresearch)
  • That makes the pattern easy to port to other measurable workflows, from model tuning to data prep to CI-style optimization.
  • It is still narrow, not recursive intelligence, but the narrowness is the whole value proposition because teams can ship it today.
// TAGS
autoresearchagentresearchautomationopen-sourcellmmlops

DISCOVERED

20d ago

2026-03-23

PUBLISHED

20d ago

2026-03-23

RELEVANCE

9/ 10

AUTHOR

tekz