BACK_TO_FEEDAICRIER_2
RTX PRO 6000 Max-Q underwhelms
OPEN_SOURCE ↗
REDDIT · REDDIT// 4h agoBENCHMARK RESULT

RTX PRO 6000 Max-Q underwhelms

A LocalLLaMA user reports unexpectedly weak llama-bench and Geekbench results from a new RTX PRO 6000 Blackwell Max-Q across both Ubuntu and Windows. Early replies point to workload choice, benchmarking method, and confusion between the 300W Max-Q card and the full 600W Workstation variant.

// ANALYSIS

This is less a clean “bad GPU” story than a useful warning about buying expensive pro AI hardware before the software stack and benchmark expectations are nailed down.

  • Max-Q keeps the same 96GB GDDR7 headline appeal, but its 300W power cap makes comparisons against the 600W RTX PRO 6000 Workstation misleading
  • LLM inference results depend heavily on llama.cpp build flags, CUDA/toolkit versions, quantization format, backend maturity, and whether Blackwell paths are optimized
  • The Windows and Ubuntu mismatch narrows the likely causes, but cross-platform slowness still does not prove defective hardware without controlled power, clocks, thermals, and known-good benchmark runs
  • Community comparisons against dual 5090s and other RTX PRO 6000 setups are valuable, but apples-to-apples model, quant, context, batch, and backend settings matter more than GPU nameplate alone
// TAGS
nvidia-rtx-pro-6000-blackwell-max-qgpuinferencebenchmarkllmself-hosted

DISCOVERED

4h ago

2026-04-21

PUBLISHED

6h ago

2026-04-21

RELEVANCE

6/ 10

AUTHOR

YouBePortnt