The worlds first AI System built on NVIDIA A100

NVIDIA DGX™ A100 is the universal system for all AI workloads, offering unprecedented compute density, performance, and flexibility in the world’s first 5 petaFLOPS AI system. NVIDIA DGX A100 features the world’s most advanced accelerator, the NVIDIA A100 Tensor Core GPU, enabling enterprises to consolidate training, inference, and analytics into a unified, easy-to-deploy AI infrastructure that includes direct access to NVIDIA AI experts.

Find out more
Video

High-Performance Computing

A100 introduces double-precision Tensor Cores, providing the biggest milestone since the introduction of double-precision computing in GPUs for HPC. This enables researchers to reduce a 10-hour, double-precision simulation running on NVIDIA V100 Tensor Core GPUs to just four hours on A100. HPC applications can also leverage TF32 precision in A100’s Tensor Cores to achieve up to 10X higher throughput for single-precision dense matrix multiply operations.

9X More HPC
Performance in 4 Years

Throughput for Top HPC Apps​

Chart 1

Up to 6X Higher Out-of-the-Box
Performance ​with TF32 for AI Training

BERT Training

Chart 2 Chart 2
  • 8X NVIDIA A100 GPUS WITH 320 GB TOTAL GPU MEMORY
    12 NVLinks/GPU, 600 GB/s GPU-to-GPU Bi-directonal Bandwidth

  • 6X NVIDIA NVSWITCHES
    4.8 TB/s Bi-directional Bandwidth, 2X More than Previous Generation NVSwitch

  • 9x MELLANOX CONNECTX-6 200Gb/S NETWORK INTERFACE
    450 GB/s Peak Bi-directional Bandwidth

  • DUAL 64-CORE AMD CPUs AND 1 TB SYSTEM MEMORY
    3.2X More Cores to Power the Most Intensive AI Jobs

  • 15 TB GEN4 NVME SSD
    25GB/s Peak Bandwidth, 2X Faster than Gen3 NVME SSDs

Essential building block of the AI data center

Icon1

The Universal System for Every AI Workload

NVIDIA DGX A100 is the universal system for all AI infrastructure, from analytics to training to inference. It sets a new bar for compute density, packing 5 petaFLOPS of AI performance into a 6U form factor, replacing legacy infrastructure silos with one platform for every AI workload.

Icon2

DGXperts: Integrated Access to AI Expertise

NVIDIA DGXperts are a global team of 14,000+ AI-fluent professionals who have built a wealth of experience over the last decade to help you maximize the value of your DGX investment.

Icon3

Fastest Time To Solution

NVIDIA DGX A100 is the world’s first AI system built on the NVIDIA A100 Tensor Core GPU. Integrating eight A100 GPUs, the system provides unprecedented acceleration and is fully optimized for NVIDIA CUDA-X™ software and the end-to-end NVIDIA data center solution stack.

Icon4

Unmatched Data Center Scalability

NVIDIA DGX A100 features Mellanox ConnectX-6 VPI HDR InfiniBand/Ethernet network adapters with 450 gigabytes per second (GB/s) of peak bi-directional bandwidth. This is one of the many features that make DGX A100 the foundational building block for large AI clusters such as NVIDIA DGX SuperPOD™, the enterprise blueprint for scalable AI infrastructure.

The worlds first Ai system built on NVIDIA A100

View:
DGX with NVIDIA A100
GPUs 8x NVIDIA A100 Tensor Core GPUs
GPU Memory 320 GB total
Perfomance 5 petaFLOPS AI
10 petaOPS INT8
NVIDIA NVSwitches 6
System Power Usage 6.5kW max
CPU Dual AMD Rome 7742, 128 cores total, 2.25 GHz (base), 3.4 GHz (max boost)
System Memory 1TB
Networking 8x Single-Port Mellanox ConnectX-6 VPI 200Gb/s HDR InfiniBand, 1x Dual-Port Mellanox ConnectX-6 VPI, 10/25/50/100/200Gb/s Ethernet
Storage OS: 2x 1.92TB M.2 NVME drives Internal Storage: 15TB (4x 3.84TB) U.2 NVME drives
Software Ubuntu Linux OS
System 271 lbs (123 kgs)
Packaged System Weight 315 lbs (143kgs)
System Dimensions Height: 10.4 in (264.0 mm), Width: 19.0 in (482.3 mm) MAX, Length: 35.3 in (897.1 mm) MAX
Operating Temperature Range 5ºC to 30ºC (41ºF to 86ºF)
Find out more Find out more