OPEN_SOURCE ↗
REDDIT · REDDIT// 4h agoBENCHMARK RESULT
RTX PRO 6000 Max-Q underwhelms
A LocalLLaMA user reports unexpectedly weak llama-bench and Geekbench results from a new RTX PRO 6000 Blackwell Max-Q across both Ubuntu and Windows. Early replies point to workload choice, benchmarking method, and confusion between the 300W Max-Q card and the full 600W Workstation variant.
// ANALYSIS
This is less a clean “bad GPU” story than a useful warning about buying expensive pro AI hardware before the software stack and benchmark expectations are nailed down.
- –Max-Q keeps the same 96GB GDDR7 headline appeal, but its 300W power cap makes comparisons against the 600W RTX PRO 6000 Workstation misleading
- –LLM inference results depend heavily on llama.cpp build flags, CUDA/toolkit versions, quantization format, backend maturity, and whether Blackwell paths are optimized
- –The Windows and Ubuntu mismatch narrows the likely causes, but cross-platform slowness still does not prove defective hardware without controlled power, clocks, thermals, and known-good benchmark runs
- –Community comparisons against dual 5090s and other RTX PRO 6000 setups are valuable, but apples-to-apples model, quant, context, batch, and backend settings matter more than GPU nameplate alone
// TAGS
nvidia-rtx-pro-6000-blackwell-max-qgpuinferencebenchmarkllmself-hosted
DISCOVERED
4h ago
2026-04-21
PUBLISHED
6h ago
2026-04-21
RELEVANCE
6/ 10
AUTHOR
YouBePortnt