OPEN_SOURCE ↗
REDDIT · REDDIT// 2h agoINFRASTRUCTURE
AMD Dual 7900 XTX Workstation Targets Local LLMs
This Reddit post asks whether a Threadripper 3960X build with 128GB RAM and two Radeon RX 7900 XTX cards is a practical local-inference machine for coding models around the 35B range. It also asks whether Linux still has the edge for ROCm, or whether AMD’s newer Windows support is good enough now.
// ANALYSIS
This is a sensible cost-driven build for local AI, but it is still a compatibility project as much as a hardware purchase. My read is that Linux remains the safer default, while Windows is no longer the non-starter it used to be.
- –AMD now officially supports the RX 7900 XTX on ROCm for Linux, and its 2025 ROCm update added a public-preview Windows PyTorch stack for Radeon 7000 and 9000 series.
- –Two 24GB cards can make 35B-class inference viable, but the experience depends heavily on whether your serving stack can split work cleanly across GPUs.
- –The 128GB system RAM is useful for host-side buffering and larger contexts, but it does not replace VRAM or create a true pooled 48GB memory space for most workloads.
- –For coding models, the practical question is less raw specs and more software maturity: model loader, quantization format, kernel support, and multi-GPU behavior will determine whether this feels fast or fragile.
- –The post reflects the current local-LLM buying reality: use what is available and affordable, then choose the OS that minimizes setup and driver pain.
// TAGS
llmgpuinferenceself-hostedamd-radeon-rx-7900-xtx
DISCOVERED
2h ago
2026-04-19
PUBLISHED
4h ago
2026-04-19
RELEVANCE
7/ 10
AUTHOR
Researchlabz