AS -8126GS-TNMR2
Featured Product Lines
Form Factor

8U rackmount

- 8 PCIe 5.0 x16 low-profile slots connected to GPU via PCIe switch
- 2 PCIe 5.0 x16 full-height full-length slots
- Optional 2 PCIe 5.0 x16 slots via expansion kit

Expansion Slots

- 12 PCIe 5.0 x4 NVMe U.2 drives
- 4 PCIe 5.0 x4 NVMe U.2 drives (optional)
- 1 M.2 NVMe/SATxA boot drive
- 2 hot-swap 2.5" SATA drives

Storage

6x or 8x 3000W N+N redundant Titanium-Level power supplies

Shared Power

- 1 RJ45 Dedicated IPMI LAN port
- 2 USB 3.0 Ports (rear)
- 1 VGA Connector

I/O Ports

- Dual SP5 sockets for AMD EPYC™ 9004 Series processors up to 400W or AMD EPYC 9005 Series processors up to 500W (two CPUs required)
- Up to 128 cores (EPYC 9004 Series) or 192 cores (EPYC 9005 Series) per CPU

Processor Support

Up to 24 DIMMs for up to 9 TB of DDR5-6000 memory

Memory Slots & Capacity
Featured System Specs

Regardless of your data center’s management approach, our open management APIs and tools are ready to support you. In addition to a dedicated IPMI port, and a Web IPMI interface, Supermicro® SuperCloud Composer software helps you configure, maintain, and monitor all of your systems using single-pane-of-glass management. If your DevOps teams prefer to use their own tools, industry-standard Redfish® APIs provide access to higher-level tools and scripting languages.

Open Management

Whatever the source of your AI platforms, AMD ROCm™ software opens doors to new levels of freedom. With support for open frameworks like PyTorch and TensorFlow, ROCm simplifies AI model migration and deployment, optimizing hardware efficiency with minimal code changes. Through strategic partnerships with AI leaders such as OpenAI, PyTorch, Hugging Face, and Databricks, the ROCm ecosystem delivers high-performance, out-of-the-box AI solutions, empowering enterprises to meet their goals with seamless integration and robust partner support.

Fast Time to Value with the AMD ROCm Platform

You can achieve faster time to results when accelerators can consume the data they need—when they need it. AMD EPYC 9005 Series processors provide up to 192 cores per CPU and up to 9 TB of memory for the parallelism you need to manage data before and/or after processing by the GPU. For tasks requiring fast per-core speed with less parallelism, the 64-core, frequency-optimized, EPYC 9575F is AI-optimized to deliver exceptional performance per core and per thread.

This AI building-block server is designed to provide each accelerator with x16 connectivity to a dedicated 400-Gbps networking device and to the host CPU—so whether data is arriving from main memory or a networked-based data lake, it can transfer directly to accelerator memory. When buffering is needed, each GPU is switched to two x8 hot-swap NVMe drive slots for a total of 16 drives dedicated to GPUs per server.

The AMD EPYC CPU’s system-on-chip (SoC) design supports built-in functions including IPMI-based management, on-board M.2 drive, and built-in SATA controllers for two drives. The SoC-oriented design reduces the number of external chip sets, helping to reduce complexity and power consumption. Titanium-Level power supplies keep the GPUs accelerating your workloads while dual-zone cooling with 10 counter-rotating fans keep the accelerators within their thermal envelopes.

Balanced System Design

The server hosts the AMD Instinct MI325X Platform, an industry-standard-based universal baseboard (UBB 2.0) model with 8 AMD Instinct MI325X accelerators and a total of 2 TB of HBM3 memory to help process the most demanding AI models. The Instinct MI325X boasts an improved HBM memory system with a 33% increase in HBM3 memory compared to the prior-generation MI300X, and 6 TB/s memory bandwidth. It is designed to hold a one-trillion parameter model in its memory. Along with a 25% improvement in FP8 throughput, it brings delivered teraFLOPS to 1565, speeding AI inference and model fine tuning. Native sparsity support helps save power, use fewer compute cycles, and reduce memory use. Each accelerator on the UBB platform connects to the other seven with 128 GB/s AMD Infinity Fabric™ Link technology for an aggregate 896 GB/s capacity. Each accelerator can connect to the host through 16 lanes of PCIe 5.0 bandwidth, and the AS -8126GS-TNMR2 is optimized for I/O throughput. Together, these features give the capacity to propel the most challenging AI workloads and large-language models.

AMD Instinct-Accelerated Server
System Highlights
Streamline Deployment at Scale for the Largest AI and Large-Language Models

Proven 8U high-performance fabric 8-GPU system design with AMD Instinct™ MI325X accelerators:

  • Industry-standard OCP accelerator module (OAM) with 8 GPUs interconnected on an AMD universal base board (UBB 2.0)

  • Industry-leading 2TB of HBM3 memory in a single server node

  • 400-Gbps networking dedicated to each GPU for large-scale AI and HPC clusters

  • 2-socket design supports 4th and 5th Gen AMD EPYC™ Processors

  • Up to 24 DIMMs for up to 9 TB of DDR5-6000 memory (with 5th Gen AMD EPYC processors)

  • Flexible PCIe 5.0 options for I/O and networking

  • Titanium-Level efficiency power supplies

  • Industrial Automation

  • Business Intelligence & Analytics

  • Large Scale Deep Learning

  • Large Language Model Training

  • AI-fused HPC applications

Key Applications:

When artificial intelligence (AI) workloads can tap into massive computational power, scientists and researchers can solve the unsolvable. Supermicro unleashes the power of large-scale infrastructure with a server built with our proven AI building-block system and powered by 5th Gen AMD EPYC™ processors and AMD Instinct™ MI325X GPU accelerators.

Next-Gen Large Scale AI Training Platform
H14 8U 8-GPU System