Home Enterprise GIGABYTE G262-ZR0 With NVIDIA HGX A100 4-GPU Released

GIGABYTE G262-ZR0 With NVIDIA HGX A100 4-GPU Released

by Adam Armstrong
GIGABYTE G262-ZR0

Today GIGABYTE Technology released a new sever for HPC, AI, and data analytics, the GIGABYTE G262-ZR0. The new server is built all around high-level performance in GPU computing. The server leverages PCIe 4.0 throughput and NVIDIA HGX as well as NVIDIA NVLink for ultra-high bandwidth performance.

Today GIGABYTE Technology released a new sever for HPC, AI, and data analytics, the GIGABYTE G262-ZR0. The new server is built all around high-level performance in GPU computing. The server leverages PCIe 4.0 throughput and NVIDIA HGX as well as NVIDIA NVLink for ultra-high bandwidth performance.

GIGABYTE G262-ZR0

The GIGABYTE G262-ZR0 is a 2U server but claims to be highly dense for GPUs. The server leverages dual second-gen AMD EPYC processors with up to 128 cores and 160 PCIe Gen 4 lanes are required for the max throughput between CPU-to-CPU and CPU-to-GPU connections. For memory, the G262 has 16 DIMM slots for a total of 4TB of DDR4, 3200MHz RAM. There are 6 low-profile (Gen 4) slots, one OCP 3.0 slot, and dual 1GbE LAN ports. For storage, the server only has six bays, four 2.5” and two M.2 slots, but GPU is more of the focus here. The server is designed with an emphasis on thermal design as all those GPUs, CPUs, and memory need to be kept cool.

Key benefits/features of the GIGABYTE G262-ZR0 include:

  • NVIDIA A100 with 40GB or 80GB: 40GB of VRAM with 1.6 TB/s of memory bandwidth or 80G of VRAM with 2.0TB/s bandwidth for high-level computational throughput.
  • Excellent GPU-to-GPU communication via 3rd gen NVIDIA NVLink with 600GB/s bandwidth.
  • Reduction in latency and CPU utilization with Mellanox Socket Direct technology. In this dual-socket server, a single CPU can access the network by bypassing the inter-processor communication bus and adjacent CPU.
  • PCIe 4.0 allows for faster interconnect (compared to PCIe 3.0) and low latency for NICs and NVMe drives via the PCIe switch fabric.
  • Ultra-fast 200Gbps access of GPUs on other servers with RDMA and HDR InfiniBand.

GIGABYTE Technology

Engage with StorageReview

Newsletter | YouTube | Podcast iTunes/Spotify | Instagram | Twitter | Facebook | RSS Feed

Advertisement
Share via
Copy link
Powered by Social Snap