BACK_TO_FEEDAICRIER_2
AMD Dual 7900 XTX Workstation Targets Local LLMs
OPEN_SOURCE ↗
REDDIT · REDDIT// 2h agoINFRASTRUCTURE

AMD Dual 7900 XTX Workstation Targets Local LLMs

This Reddit post asks whether a Threadripper 3960X build with 128GB RAM and two Radeon RX 7900 XTX cards is a practical local-inference machine for coding models around the 35B range. It also asks whether Linux still has the edge for ROCm, or whether AMD’s newer Windows support is good enough now.

// ANALYSIS

This is a sensible cost-driven build for local AI, but it is still a compatibility project as much as a hardware purchase. My read is that Linux remains the safer default, while Windows is no longer the non-starter it used to be.

  • AMD now officially supports the RX 7900 XTX on ROCm for Linux, and its 2025 ROCm update added a public-preview Windows PyTorch stack for Radeon 7000 and 9000 series.
  • Two 24GB cards can make 35B-class inference viable, but the experience depends heavily on whether your serving stack can split work cleanly across GPUs.
  • The 128GB system RAM is useful for host-side buffering and larger contexts, but it does not replace VRAM or create a true pooled 48GB memory space for most workloads.
  • For coding models, the practical question is less raw specs and more software maturity: model loader, quantization format, kernel support, and multi-GPU behavior will determine whether this feels fast or fragile.
  • The post reflects the current local-LLM buying reality: use what is available and affordable, then choose the OS that minimizes setup and driver pain.
// TAGS
llmgpuinferenceself-hostedamd-radeon-rx-7900-xtx

DISCOVERED

2h ago

2026-04-19

PUBLISHED

4h ago

2026-04-19

RELEVANCE

7/ 10

AUTHOR

Researchlabz