ARS-121L-DNR
1U 2-Node NVIDIA Grace CPU Superchip GPU Server supporting NVIDIA BlueField-3 or ConnectX-7
Two nodes in a 1U form factor. Each node supports the following:
- High density 1U 2-node system with NVIDIA Grace™ CPU Superchip per node
- NVIDIA Grace™ CPU Superchip (144-core per node)
- NVLink® Chip-2-Chip (C2C) high-bandwidth, low-latency interconnect between CPU and CPU at 900GB/s
- Up to 480GB LPDDR5X onboard memory
- 2x PCIe 5.0 x16 slots per node supporting NVIDIA BlueField®-3 or ConnectX®-7
- Up to 4x Hot-swap E1.S drives and 2x M.2 NVMe drives per node
- 7 Hot-Swap Heavy Duty Fans with Optimal Fan Speed Control
Key Applications
- High Performance Computing
- Hyperscale Cloud Applications
- Data Analytics
Product Specification
| Product SKUs | ARS-121L-DNR (Silver) |
| Motherboard | Super G1SMH |
| Processor (per Node) | |
| CPU | Single processor(s) NVIDIA Dual 72-core CPUs on a Grace™ CPU Superchip |
| Note | Supports up to 500W TDP CPUs (Air Cooled) |
| GPU (per Node) | |
| Max GPU Count | Up to 1 double-width or 1 single-width GPU(s) |
| GPU-GPU Interconnect | PCIe |
| System Memory (per Node) | |
| Memory | Slot Count: Onboard Memory Max Memory: Up to 480GB ECC |
| On-Board Devices (per Node) | |
| Chipset | System on Chip |
| Network Connectivity | 1x 1GbE BaseT with NVIDIA ConnectX®-7 or Bluefield®-3 DPU |
| Input / Output (per Node) | |
| LAN | 1 RJ45 1GbE (Dedicated IPMI port) |
| System BIOS | |
| BIOS Type | AMI 32MB SPI Flash EEPROM |
| PC Health Monitoring | |
| CPU | 8+4 Phase-switching voltage regulator Monitors for CPU Cores, Chipset Voltages, Memory |
| FAN | Fans with tachometer monitoring Pulse Width Modulated (PWM) fan connectors Status monitor for speed control |
| Temperature | Monitoring for CPU and chassis environment Thermal Control for fan connectors |
| Chassis | |
| Form Factor | 1U Rackmount |
| Model | CSE-GP102TS-R000NDFP |
| Dimensions and Weight | |
| Height | 1.75" (44mm) |
| Width | 17.33" (440mm) |
| Depth | 37" (940mm) |
| Package | 9.5" (H) x 48" (W) x 28" (D) |
| Weight | Net Weight: 48.5 lbs (22 kg) Gross Weight: 65.5 lbs (29.7 kg) |
| Available Color | Silver |
| Expansion Slots (per Node) | |
| PCI-Express (PCIe) | 2 PCIe 5.0 x16 FHFL slot(s) |
| Drive Bays / Storage (per Node) | |
| Hot-swap | 4x E1.S hot-swap NVMe drive slots |
| M.2 | 2 M.2 NVMe |
| System Cooling | |
| Fans | 7 Removable heavy-duty 4CM Fan(s) |
| Power Supply | 2x 2700W Redundant Titanium Level power supplies |
| Operating Environment | |
| Environmental Spec. | Operating Temperature: 10°C ~ 35°C (50°F ~ 95°F) Non-operating Temperature: -40°C to 60°C (-40°F to 140°F) Operating Relative Humidity: 8% to 90% (non-condensing) Non-operating Relative Humidity: 5% to 95% (non-condensing) |
Grace and x86 MGX System Configurations at a Glance
Supermicro NVIDIA MGX™ 1U/2U Systems with Grace™ CPU Superchip and x86 CPUs are fully optimized to support up to 4 GPUs via PCle without sacrificing I/O networking, or thermals. The ultimate building block architecture allows you to tailor these systems optimized for a variety of accelerated workloads and fields, including Al training and inference, HPC, data analytics, visualization/Omniverse™, and hyperscale cloud applications.
| SKU | ARS-121L-DNR | ARS-221GL-NR | SYS-221GE-NR |
![]() | ![]() | ![]() | |
| Form Factor | 1U 2-node system with NVIDIA Grace CPU Superchip per node | 2U GPU system with single NVIDIA Grace CPU Superchip | 2U GPU system with dual x86 CPUs |
| CPU | 144-core Grace Arm Neoverse V2 CPU in a single chip per node (total of 288 cores in one system) | 144-core Grace Arm Neoverse V2 CPU in a single chip | 4th Gen Intel Xeon Scalable Processors (Up to 56-core per socket) |
| GPU | Please contact our sales for possible configurations | Up to 4 double-width GPUs including NVIDIA H100 PCIe, H100 NVL, L40S | Up to 4 double-width GPUs including NVIDIA H100 PCIe, H100 NVL, L40S |
| Memory | Up to 480GB of integrated LPDDR5X memory with ECC and up to 1TB/s of bandwidth per node | Up to 480GB of integrated LPDDR5X memory with ECC and up to 1TB/s of bandwidth per node | Up to 2TB, 32x DIMM slots, ECC DDR5-4800 |
| Drives | Up to 4x hot-swap E1.S drives and 2x M.2 NVMe drives per node | Up to 8x hot-swap E1.S drives and 2x M.2 NVMe drives | Up to 8x hot-swap E1.S drives and 2x M.2 NVMe drives |
| Networking | 2x PCIe 5.0 x16 slots per node supporting NVIDIA BlueField-3 or ConnectX-7 (e.g., 1 GPU and 1 BlueField-3) | 3x PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or ConnectX-7 (in addition to 4x PCIe 5.0 x16 slots for GPUs) | 3x PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or ConnectX-7 (in addition to 4x PCIe 5.0 x16 slots for GPUs) |
| Interconnect | NVLink™-C2C with 900GB/s for CPU-CPU interconnect (within node) | NVLink Bridge GPU-GPU interconnect supported (e.g., H100 NVL) | NVLink™ Bridge GPU-GPU interconnect supported (e.g., H100 NVL) |
| Cooling | Air-cooling | Air-cooling | Air-cooling |
| Power | 2x 2700W Redundant Titanium Level power supplies | 3x 2000W Redundant Titanium Level power supplies | 3x 2000W Redundant Titanium Level power supplies |




