NVIDIA Hopper & Ampere AI GPUs Continue To Post World Records In AI Training Benchmarks

Hassan Mujtaba

NVIDIA has just published even more world records in AI Training benchmarks with its powerful Hopper and Ampere GPUs.

NVIDIA Continues To Sweep AI Training Benchmarks With Hopper & Ampere GPUs

Press Release: In industry-standard tests of AI training, NVIDIA H100 Tensor Core GPUs set world records on enterprise workloads; A100 raised the bar in high-performance computing.

Related Story Elon Musk Says That Tesla’s AI Training Capacity Will Be Equivalent to Around 85,000 Units of NVIDIA’s H100 Chips by the End of 2024
nvidia-ai-training-benchmarks-world-record-hopper-ampere-gpus-_1-2
nvidia-ai-training-benchmarks-world-record-hopper-ampere-gpus-_2-2
nvidia-ai-training-benchmarks-world-record-hopper-ampere-gpus-_3-2
nvidia-ai-training-benchmarks-world-record-hopper-ampere-gpus-_7
nvidia-ai-training-benchmarks-world-record-hopper-ampere-gpus-_8
nvidia-ai-training-benchmarks-world-record-hopper-ampere-gpus-_10

Two months after their debut sweeping MLPerf inference benchmarks, NVIDIA H100 Tensor Core GPUs set world records across enterprise AI workloads in the industry group’s latest tests of AI training. Together, the results show H100 is the best choice for users who demand the utmost performance when creating and deploying advanced AI models.

MLPerf is the industry standard for measuring AI performance. It’s backed by a broad group that includes Amazon, Arm, Baidu, Google, Harvard University, Intel, Meta, Microsoft, Stanford University, and the University of Toronto. In a related MLPerf benchmark also released today, NVIDIA A100 Tensor Core GPUs raised the bar they set last year in high-performance computing (HPC).

NVIDIA H100 GPUs (aka Hopper) set world records for training models in all eight MLPerf enterprise workloads. They delivered up to 6.7x more performance than previous-generation GPUs when they were first submitted on MLPerf training. By the same comparison, today’s A100 GPUs pack 2.5x more muscle, thanks to advances in software.

A100 GPUs Hit New Peak in HPC

In the separate suite of MLPerf HPC benchmarks, A100 GPUs swept all tests of training AI models in demanding scientific workloads run on supercomputers. The results show the NVIDIA AI platform’s ability to scale to the world’s toughest technical challenges.

For example, A100 GPUs trained AI models in the CosmoFlow test for astrophysics 9x faster than the best results two years ago in the first round of MLPerf HPC. In that same workload, the A100 also delivered up to a whopping 66x more throughput per chip than an alternative offering.

The HPC benchmarks train models for work in astrophysics, weather forecasting, and molecular dynamics. They are among many technical fields, like drug discovery, and adopting AI to advance science.

Supercomputer centers in Asia, Europe, and the U.S. participated in the latest round of the MLPerf HPC tests. In its debut on the DeepCAM benchmarks, Dell Technologies showed strong results using NVIDIA A100 GPUs.

An Unparalleled Ecosystem

In the enterprise AI training benchmarks, a total of 11 partners, including the Microsoft Azure cloud service, made submissions using NVIDIA A100, A30, and A40 GPUs. System makers including ASUS, Dell Technologies, Fujitsu, GIGABYTE, Hewlett Packard Enterprise, Lenovo, and Supermicro used a total of nine NVIDIA-Certified Systems for their submissions.

For example, submissions in the latest HPC tests applied a suite of software optimizations and techniques described in a technical article. Together they slashed runtime on one benchmark by 5x, to just 22 minutes from 101 minutes.

In the latest round, at least three partners joined NVIDIA in submitting results on all eight MLPerf training workloads. That versatility is important because real-world applications often require a suite of diverse AI models.

Share this story

Deal of the Day

Comments