OPEN_SOURCE ↗
REDDIT · REDDIT// 32d agoBENCHMARK RESULT
Autoresearch-ANE slashes val_loss on Mac
The autoresearch-ane fork of Karpathy's autoresearch reports its best Apple Neural Engine run yet, cutting validation loss from 6.109 to 3.55 on an M3 MacBook. The big unlock was a dynamic weight pipeline that avoids constant recompilation and reportedly delivers about 11x more training steps in the same 5-minute budget.
// ANALYSIS
This is a meaningful systems result, not just a prettier loss curve: the project gets much more real training done on consumer Apple hardware once compilation stops dominating the wall clock.
- –The repo says the dynamic pipeline compiles 10 ANE kernels once at startup, then stages weights dynamically, boosting throughput from roughly 120 to about 1340 steps per 5-minute run
- –The ANE backend is its own training stack in Objective-C with TinyStories and `val_loss`, so the numbers are impressive but not directly comparable to Karpathy's CUDA `val_bpb` baseline
- –Keeping the agent's edit surface mostly to `ane/experiment_config.h` makes autonomous overnight experimentation much more plausible on a laptop
- –If these gains hold up, Apple's Neural Engine starts looking less like an inference novelty and more like a viable playground for small-model research loops
// TAGS
autoresearch-aneagentllmopen-sourceedge-aibenchmark
DISCOVERED
32d ago
2026-03-11
PUBLISHED
32d ago
2026-03-11
RELEVANCE
8/ 10
AUTHOR
paraboloed