BACK_TO_FEEDAICRIER_2
preflight catches PyTorch training bugs before run
OPEN_SOURCE ↗
REDDIT · REDDIT// 27d agoOPENSOURCE RELEASE

preflight catches PyTorch training bugs before run

preflight-ml is a CLI pre-training validator for PyTorch that runs 10 safety checks — including label leakage, NaN detection, gradient issues, and VRAM estimation — in ~30 seconds before training starts. Built by a developer who lost three days to silent label leakage, it exits with code 1 on fatal failures to block bad runs in CI.

// ANALYSIS

Silent training failures are one of the most painful failure modes in ML, and preflight-ml is a simple, opinionated answer to a gap that bigger tools like Deepchecks don't neatly fill.

  • Ten checks across fatal/warn/info severity tiers catch the silent bugs (NaNs, label leakage, wrong channel ordering) that waste GPU hours without ever throwing an error
  • CI integration via exit codes is the right call — this belongs in the pre-training pipeline, not as a post-hoc debugging tool
  • Operates on sampled batches (~30 seconds), so it's fast enough to be a default step without adding meaningful overhead
  • No direct competitor occupies this exact niche: Deepchecks is heavier, Great Expectations is data-only, PyTorch Lightning embeds similar checks but only inside its own framework
  • v0.1.1 alpha with a clear roadmap (drift detection, auto-patching, domain plugins) — early but genuinely useful even now
// TAGS
preflight-mldevtoolmlopsopen-sourceclitesting

DISCOVERED

27d ago

2026-03-15

PUBLISHED

27d ago

2026-03-15

RELEVANCE

7/ 10

AUTHOR

Red_Egnival