BACK_TO_FEEDAICRIER_2
Supermicro V100 rig eyes local LLMs
OPEN_SOURCE ↗
REDDIT · REDDIT// 21d agoINFRASTRUCTURE

Supermicro V100 rig eyes local LLMs

A Reddit user is eyeing a dirt-cheap 8x Tesla V100 setup for local LLMs, pairing used datacenter GPUs with an older Supermicro chassis and custom water cooling. The big catch is the chassis choice: the official 8x V100 NVLink platform is the SYS-4028GR-TVRT, while TXRT/TRT point at different GPU layouts and generations.

// ANALYSIS

This is a clever salvage-build idea, but it is much more likely to become a fun hardware project than a clean cost/perf king.

  • Supermicro’s own docs show the SYS-4028GR-TVRT as the 8x Tesla V100 SXM2, 300 GB/s NVLink box; TXRT is the P100-era sibling, and TRT is a PCIe-gpu chassis.
  • If you end up with PCIe V100s instead of SXM2 modules, the whole NVLink premise changes, so the exact GPU form factor matters as much as the price.
  • 128 GB of aggregate VRAM sounds huge, but it is still sharded memory across eight cards, so model parallelism, interconnect efficiency, and software support will decide real-world speed.
  • Custom water cooling can make the thermals work, but it also adds leak risk, maintenance headaches, and another layer of failure on top of already old enterprise hardware.
  • The value case is strongest if you want a tinkering lab and can tolerate rough edges; if you want a low-friction local inference box, simpler modern hardware is usually the saner buy.
// TAGS
supermicro-sys-4028gr-tvrttesla-v100llmgpuinferenceself-hostednvlink

DISCOVERED

21d ago

2026-03-21

PUBLISHED

21d ago

2026-03-21

RELEVANCE

8/ 10

AUTHOR

lethalratpoison