OPEN_SOURCE ↗
REDDIT · REDDIT// 2d agoINFRASTRUCTURE
Cloudflare R2 jitters starve H100s
A Machine Learning subreddit post says Cloudflare R2’s zero-egress pitch breaks down under AI training load because TTFB jitter leaves H100s waiting on I/O. The poster wants an S3-compatible store that can stream 40TB-scale datasets without forcing a custom NVMe cache layer.
// ANALYSIS
The sharp take: zero egress is not the same as training-ready storage. For GPU-heavy workloads, predictable latency and locality matter more than headline storage pricing.
- –The post highlights a common infra trap: object storage that is cheap on paper can still waste expensive GPU time if read latency is noisy.
- –Alternatives like Wasabi and Backblaze B2 can remove egress fees too, but they come with their own pricing or policy tradeoffs, so “free egress” is not a complete answer.
- –At 40TB scale, the likely winning pattern is regional replication plus a local cache tier, not direct-from-object-store streaming as the only data path.
- –This is less a Cloudflare-specific complaint than a reminder that AI training pipelines need storage engineered for throughput consistency, not just S3 compatibility.
// TAGS
cloudgpudata-toolsmlopscloudflare-r2
DISCOVERED
2d ago
2026-04-09
PUBLISHED
3d ago
2026-04-09
RELEVANCE
7/ 10
AUTHOR
regentwells