OPEN_SOURCE ↗
REDDIT · REDDIT// 27d agoOPENSOURCE RELEASE
preflight catches PyTorch training bugs before run
preflight-ml is a CLI pre-training validator for PyTorch that runs 10 safety checks — including label leakage, NaN detection, gradient issues, and VRAM estimation — in ~30 seconds before training starts. Built by a developer who lost three days to silent label leakage, it exits with code 1 on fatal failures to block bad runs in CI.
// ANALYSIS
Silent training failures are one of the most painful failure modes in ML, and preflight-ml is a simple, opinionated answer to a gap that bigger tools like Deepchecks don't neatly fill.
- –Ten checks across fatal/warn/info severity tiers catch the silent bugs (NaNs, label leakage, wrong channel ordering) that waste GPU hours without ever throwing an error
- –CI integration via exit codes is the right call — this belongs in the pre-training pipeline, not as a post-hoc debugging tool
- –Operates on sampled batches (~30 seconds), so it's fast enough to be a default step without adding meaningful overhead
- –No direct competitor occupies this exact niche: Deepchecks is heavier, Great Expectations is data-only, PyTorch Lightning embeds similar checks but only inside its own framework
- –v0.1.1 alpha with a clear roadmap (drift detection, auto-patching, domain plugins) — early but genuinely useful even now
// TAGS
preflight-mldevtoolmlopsopen-sourceclitesting
DISCOVERED
27d ago
2026-03-15
PUBLISHED
27d ago
2026-03-15
RELEVANCE
7/ 10
AUTHOR
Red_Egnival