BACK_TO_FEEDAICRIER_2
7900 XTX Multi-GPU Support Still Bumpy
OPEN_SOURCE ↗
REDDIT · REDDIT// 3h agoINFRASTRUCTURE

7900 XTX Multi-GPU Support Still Bumpy

ROCm and vLLM now officially support the Radeon RX 7900 series, and vLLM exposes tensor, pipeline, and data parallelism on AMD GPUs. But dual-7900 XTX multi-GPU inference is still not turnkey: recent ROCm/RCCL reports show collective failures on 2x 7900 XTX, even outside vLLM.

// ANALYSIS

The short version: yes, tensor parallelism exists now, but the software stack is still the limiting factor, not the idea of splitting a model across two 7900 XTX cards.

  • vLLM’s current docs list Radeon RX 7900 series support on ROCm 6.3+ and document tensor parallelism as a normal multi-GPU mode
  • AMD’s ROCm docs still warn that two 7900 XTXs need proper CPU-controlled PCIe slots; chipset-attached cards can trigger HIP errors in AI workloads
  • A March 2026 ROCm issue reports dual-GPU collectives failing on 2x 7900 XTX with ROCm 7.2.1/RCCL 2.27.7, and the reporter hit it while trying to run vLLM tensor parallelism
  • vLLM has continued to improve ROCm support, but the presence of open 7900 XTX-specific bugs means consumer RDNA3 is still more "hands-on lab setup" than "drop-in production box"
  • llama.cpp also now marks parallel multi-GPU support on ROCm, which is a real improvement, but it does not remove topology and collective-communication caveats
// TAGS
gpuinferencevllmllama-cppamd-radeon-rx-7900-xtxrocm

DISCOVERED

3h ago

2026-05-01

PUBLISHED

3h ago

2026-05-01

RELEVANCE

8/ 10

AUTHOR

ziphnor