BACK_TO_FEEDAICRIER_2
Parameter Golf fits 24M LLM in 15MB
OPEN_SOURCE ↗
REDDIT · REDDIT// 18d agoBENCHMARK RESULT

Parameter Golf fits 24M LLM in 15MB

The post reports a top-3 finish in OpenAI's Parameter Golf challenge by squeezing a 24M-parameter LLM into about 15MB. The main gain comes from per-row INT8 calibration that tries five clip percentiles and keeps the lowest-MSE reconstruction, while a wider model appears to scale better than extra depth under the same cap.

// ANALYSIS

At this scale, quantization is the product, not a finishing touch. The submission reads like GPTQ-lite plus architecture tuning, which is exactly the kind of compression-first work that moves a tiny-model leaderboard.

  • Five clip candidates per row is a very cheap search space, but it directly attacks reconstruction error instead of assuming one clipping heuristic will be good enough.
  • The width-over-depth result suggests that once the byte ceiling is tight, representational breadth matters more than stacking layers.
  • Landing around 15.06MB means the margin is microscopic, so schedule tweaks, optimizer choices, and packing format can all move the ranking.
  • Because OpenAI's rules fix the dataset and cap training time, leaderboard separation is likely coming from many small compounding improvements, not one dramatic breakthrough.
// TAGS
parameter-golfllmbenchmarkresearch

DISCOVERED

18d ago

2026-03-25

PUBLISHED

18d ago

2026-03-25

RELEVANCE

9/ 10

AUTHOR

TrashFun5286