Menu

NVIDIA H100 MLPerf 2.1 AI Performance Results Are Impressive

The News: NVIDIA H100 Tensor Core GPUs, the company’s newest and most powerful chips that are used for AI and machine learning, just set new performance records in the latest round of MLPerf Inference v2.1 industry-standard benchmark tests for AI inferencing performance. In the newest round of MLPerf Inference v2.1 tests, which produce benchmarks for data center and edge uses, the NVIDIA H100 GPUs delivered up to 4.5x more performance than NVIDIA’s previous A100 GPUs. Read the full NVIDIA blog post on its latest MLPerf v2.1 AI performance test results.

NVIDIA H100 MLPerf 2.1 AI Performance Results Are Impressive

Analyst Take: With its latest NVIDIA H100 GPUs, NVIDIA again establishes itself as the company to beat when it comes to GPU performance for AI inferencing, based on the newest MLPerf Inference v2.1 benchmarks from the open source engineering consortium, MLCommons.

I realize the repetitive nature of my commentary here, but the reality is every time we write about the latest MLPerf testing results, NVIDIA is at the top of the results mix with its then-latest powerful GPUs. And based on NVIDIA’s continuing technical leadership in GPUs used for AI and ML, this should not be a surprise.

What is also impressive is that unlike competitors, NVIDIA enters its GPUs in each of the MLPerf data center and edge computing tests, rather than just hand-picking the tests where its GPUs will perform best. This gives a clear picture of the performance of NVIDIA GPUs and how they stack up against competitors in any category.

The MLPerf v2.1 test results are important in the industry because they provide an excellent way to evaluate AI inferencing performance when comparing GPUs, placing each one head-to-head while starting from an equal footing. The MLPerf testing is a respected yardstick where each vendor can submit their chips and be compared to establish real-world performance leadership.

In this latest round of MLPerf benchmark testing, the NVIDIA H100 Tensor Core GPUs set world records in inference on all workloads, delivering up to 4.5x more performance than previous-generation NVIDIA A100 Ampere GPUs. That is impressive and highlights the H100 GPUs as the chip to use to get the highest performance in AI inferencing using advanced AI models.

These were the first MLPerf tests to include the NVIDIA H100 GPUs, which are slated to be available later in 2022.

Notably, the NVIDIA H100 GPUs particularly excelled on the MLPerf benchmark test involving the BERT model for natural language processing, which is one of the largest and most performance-hungry of the MLPerf AI models.

How the NVIDIA H100 Helps Boost MLPerf Results

The dramatic performance boost for the NVIDIA H100 GPUs in the MLPerf Inference v2.1 benchmark tests comes through rich design and engineering based on decades of deep experience in GPU architecture. The H100 GPUs include 80 billion transistors and are built on a TSMC 4nm process. The NVIDIA H100 also includes a new Transformer Engine that is as much as 6x faster than previous versions, as well as a highly scalable, super-fast NVIDIA NVLink interconnect.

The NVIDIA H100 Tensor core GPU is the company’s first GPU based on its latest NVIDIA Hopper accelerated computing platform architecture that was unveiled in 2021 to replace its now three-year-old Ampere architecture.

The latest MLPerf Inference v.2.1 benchmarks H100 raised the bar in per-accelerator performance across the six neural networks that were included in the latest benchmark round. The H100 results performed admirably in both throughput and speed in separate server and offline scenarios in the benchmark testing in data center and edge scenarios,

MLPerf is a consortium of AI leaders from academia, research labs, and industry which builds what it calls “fair and useful benchmarks” that aim to produce unbiased evaluations of training and inference performance for hardware, software, and services under prescribed conditions. MLPerf does its testing at regular intervals and adds new workloads as needed to represent the state of the art in AI, according to the group. The benchmark suites are open source and peer reviewed.

The latest MLPerf Inference v2.1 tests are built to stress machine learning models, software, and hardware, while also monitoring energy consumption.

Included in this round of testing for datacenter and edge systems were entries from NVIDIA and a wide range of other hardware vendors, including Alibaba, ASUSTeK, Azure, Biren, Dell, Fujitsu, GIGABYTE, H3C, HPE, Inspur, Intel, Krai, Lenovo, Moffett, Nettrix, Neural Magic, OctoML, Qualcomm Technologies, Inc., SAPEON, and Supermicro.

Earlier NVIDIA A100 GPUs Also Continue Their Performance Leadership

The MLPerf Inference v2.1 benchmark results also showed continuing AI inferencing performance leadership in today’s marketplace for NVIDIA’s previously established A100 GPUs. In the tests, NVIDIA A100 GPUs won more tests in data center and edge computing categories and scenarios than any other competitors. NVIDIA A100 performance has been increased by 6x in the MLPerf tests since the GPUs were first tested by the organization in July of 2020.

NVIDIA MLPerf v2.1 Performance Overview

In the ultra-competitive world of AI inferencing, NVIDIA continues to display its engineering prowess and technological expertise when it comes to producing hardware and software that is built to deliver in the marketplace and on MLPerf benchmark test results.
And the latest MLPerf Inference v2.1 results show that NVIDIA is maintaining that position of competitiveness and leadership in a field that is always demanding higher performance and broader capabilities.

We believe that as NVIDIA continues to focus on the future of AI, it will continue to play a leadership role in delivering the promise and performance of AI to the world.

NVIDIA continues to be an exciting company to watch in the AI marketplace around the globe and it will be fascinating to track its next moves in this always evolving field.

Disclosure: Futurum Research is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of Futurum Research as a whole.

Other insights from Futurum Research:

NVIDIA Q2 Revenue Up 3% YoY as Gaming Slows Post-Pandemic

Intel Fab Co-Investment, Zoom Perspectives, Salesforce, NVIDIA, Dell Tech & Marvell Earnings – The Six Five Webcast

NVIDIA Omniverse Gets Expanded Tools, Frameworks, Plug-ins

Image Credit: NVIDIA
Related Insights
Is Autonomous IT the Endgame for AI in Operations or Just the Start of a Bigger Shift?
April 11, 2026

Is Autonomous IT the Endgame for AI in Operations or Just the Start of a Bigger Shift?

As Autonomous IT evolves, CIOs must weigh efficiency gains against vendor lock-in and skills gaps, raising the question: is this AI's operational endgame or just the beginning?...
Agentic AI
April 10, 2026

Oracle’s Fusion Agentic Apps: Can Platform-First AI Finally Deliver Enterprise ROI?

Oracle launches Fusion Agentic Applications with autonomous AI agents in enterprise platforms. Research shows 38.8% of enterprise buyers now expect GenAI delivery via agents, signaling a fundamental shift in how...
Can AI Save the Mainframe BMC Bets on Governance and Agentic AI
April 10, 2026

Can AI Save the Mainframe? BMC Bets on Governance and Agentic AI

Brad Shimmin and Mitch Ashley, Analysts at Futurum, examine BMC Software’s April 2026 AI expansion. The report details how uniting AMI with Control-M's new Agent Gateway addresses the mainframe demographic...
Anthropic's Gigawatt-Scale TPU Deal with Broadcom Creates a Structural Advantage
April 9, 2026

Anthropic’s Gigawatt-Scale TPU Deal with Broadcom Creates a Structural Advantage

Brendan Burke, Research Director at Futurum, examines Anthropic TPU expansion with Google and Broadcom, highlighting how multi-gigawatt compute deals and custom silicon are reshaping AI infrastructure scale and competition....
Slack Expands Slackbot for Enterprise Work; Can It Simplify Execution?
April 9, 2026

Slack Expands Slackbot for Enterprise Work; Can It Simplify Execution?

Keith Kirkpatrick, VP and Research Director at Futurum, examines Salesforce’s Slackbot enterprise update, expanding Slack into a unified work interface with AI skills, CRM, and orchestration capabilities....
Does Honoring Matei Zaharia Signal a New Era for Open-Source Data and AI Systems?
April 9, 2026

Does Honoring Matei Zaharia Signal a New Era for Open-Source Data and AI Systems?

Matei Zaharia's ACM Prize for Apache Spark reflects enterprise AI's shift toward open-source platforms, showing how democratized data infrastructure is transforming competitive dynamics across the industry....

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.