OPEN_SOURCE ↗
REDDIT · REDDIT// 29d agoINFRASTRUCTURE
ML engineers debate Docker, uv for CUDA
A Reddit discussion asks how to manage conflicting CUDA and Python dependencies across multiple ML projects without Conda pain. The proposed workflow is to pin OS/CUDA with Docker and manage Python packages with uv inside each container for faster, reproducible environments.
// ANALYSIS
Docker plus uv is a pragmatic modern default for multi-project ML work, while Conda remains a fallback when binary compatibility gets messy.
- –Containers isolate CUDA runtime, system libraries, and distro quirks better than per-project host installs.
- –uv speeds Python dependency resolution and lockfile workflows, reducing environment drift inside images.
- –NVIDIA base images and pinned tags make reproducibility explicit across teammates and CI.
- –Conda or micromamba still helps for edge cases where compiled ML packages are unavailable or fragile in pip-only setups.
// TAGS
cudadockeruvcondagpumlops
DISCOVERED
29d ago
2026-03-14
PUBLISHED
30d ago
2026-03-12
RELEVANCE
8/ 10
AUTHOR
sounthan1