Supermicro AS-8126GS-TNMR | 8x AMD Instinct MI350X GPU Server

$387,127.13

The Supermicro AS-8126GS-TNMR is an air-cooled 8U GPU server featuring 8 AMD Instinct MI350X OAM accelerators with 2.3TB total HBM3e memory, dual AMD EPYC 9575F processors (128 cores total), 2.25TB DDR5-6400 ECC memory, and 8x 400 Gb/s NDR InfiniBand/Ethernet OSFP networking via NVIDIA ConnectX-7. Purpose-built for large-scale AI training, LLM inference, generative AI, and HPC workloads requiring massive GPU memory capacity and high-bandwidth interconnect.

GPU 8x AMD Instinct MI350X 288GB HBM3e (OAM, UBB 2.0)
CPU Dual AMD EPYC 9575F (64-Core, 3.3 GHz Base / 5.0 GHz Boost, 256MB L3 Cache, 400W TDP)
RAM 2.25TB DDR5-6400 ECC RDIMM (24x 96GB)
Storage 1x 960GB PCIe Gen4 M.2 NVMe (OS); 8x 3.8TB PCIe Gen4 2.5" NVMe (30.4TB Raw)
Cooling Air Cooled
Power 6x 5250W
Network 8x 400 Gb/s OSFP (NDR InfiniBand / 400GbE) via NVIDIA ConnectX-7 PCIe Gen5 x16; 1x Mellanox ConnectX-6 Dx 100GbE QSFP56 2-Port PCIe Gen4 x16; 2x 10GbE RJ45 (Intel X710) + 1GbE IPMI/BMC
Applications AI Training, LLM, Generative AI
Chassis 8U Rackmount

Description

Accelerate your most demanding AI training, inference, and HPC workloads with the Supermicro AS-8126GS-TNMR, an 8U GPU server built on the AMD Instinct MI350X platform. Powered by 8 AMD Instinct MI350X OAM accelerators on a UBB 2.0 baseboard, dual AMD EPYC 9575F processors (64 cores, 5.0 GHz boost), and 2.25TB of DDR5-6400 ECC memory, this system delivers exceptional compute density for large-scale AI model training, LLM inference, generative AI, and scientific computing. Each MI350X GPU features 288GB of HBM3e memory with 8 TB/s bandwidth, providing a total of 2.3TB of coherent GPU memory across the platform — enough to run 520B+ parameter models without complex sharding.

The AS-8126GS-TNMR provides high-bandwidth GPU-to-GPU communication via 4th Gen AMD Infinity Fabric with a fully meshed 8-GPU topology, delivering 7x 153.6 GB/s bidirectional links per GPU for fast, low-latency data movement across all accelerators. For multi-node scaling, 8 NVIDIA ConnectX-7 adapters deliver 8x 400 Gb/s NDR InfiniBand or Ethernet connectivity via OSFP ports over PCIe Gen5 x16, enabling seamless integration into large-scale AI fabrics and supercomputing clusters with 1:1 GPU-to-network ratio. An additional Mellanox ConnectX-6 Dx dual-port 100GbE QSFP56 NIC handles dedicated management and storage traffic, while onboard dual 10GbE RJ45 ports and a dedicated BMC/IPMI port provide out-of-band management.

With 30.4TB of raw NVMe storage across 8 hot-swap 2.5″ PCIe Gen4 bays, a 960GB M.2 NVMe boot drive, 6x 5,250W Titanium-efficiency power supplies in a 3+3 redundant configuration, and 10 PCIe 5.0 expansion slots, the AS-8126GS-TNMR provides a dense, air-cooled, and scalable foundation for production AI infrastructure. Backed by a 3-year limited warranty with 3-year onsite next-business-day service, this system is ready to deploy in AI factories, neoclouds, research labs, and enterprise data centers.

ApplicationsAI Training, LLM, Generative AI
Chassis8U Rackmount
CoolingAir Cooled
CPUDual AMD EPYC 9575F (64-Core, 3.3 GHz Base / 5.0 GHz Boost, 256MB L3 Cache, 400W TDP)
Drive1x 960GB PCIe Gen4 M.2 NVMe (OS); 8x 3.8TB PCIe Gen4 2.5" NVMe (30.4TB Raw)
GPU8x AMD Instinct MI350X 288GB HBM3e (OAM, UBB 2.0)
Network Ports8x 400 Gb/s OSFP (NDR InfiniBand / 400GbE) via NVIDIA ConnectX-7 PCIe Gen5 x16; 1x Mellanox ConnectX-6 Dx 100GbE QSFP56 2-Port PCIe Gen4 x16; 2x 10GbE RJ45 (Intel X710) + 1GbE IPMI/BMC
RAM2.25TB DDR5-6400 ECC RDIMM (24x 96GB)
Power Supply6x 5250W
Storage30.4TB Raw NVMe; 8x 2.5" Hot-Swap NVMe Bays + 1x M.2 NVMe Boot