On-Prem

HPC

Nvidia claims 'record performance' for Hopper MLPerf debut

H100 Tensor Core GPU leaves A100 in the dust, but company says previous gen has improved too


Nvidia's Hopper GPU has turned in its first scores in the newly released MLPerf Inference v2.1 benchmark results, with the company claiming new records for its performance.

MLCommons analyzes the performance of systems performing inferencing tasks using a machine learning model against new data. The results are available here.

Nvidia tends to dominate the results, and this time the company said that its Hopper-based H100 Tensor Core GPUs set new records in inferencing workloads, claiming it delivers up to 4.5x more performance than previous GPUs.

The tests are effectively the first public demonstration of the H100 GPUs, which are set to be available later this year..

In this case, the previous GPU referred to appears to mean Nvidia's A100 product, which has been widely deployed in many AI and HPC systems over the past year or so. Hopper delivered improved per-accelerator performance across all six neural network models, and Nvidia claimed it represents the leadership position for both throughput and speed in separate server and offline scenarios.

However, while the A100 products may no longer be Nvidia's hottest AI platform, the company said these have continued to show gains in performance thanks to continuous improvements in Nvidia's AI software. It claimed the MLPerf figures have advanced by 6x since the A100 was first listed in the results two years ago.

Nvidia also submitted results for its Orin edge computing platform, which integrates Arm cores with an Ampere architecture GPU. The company claimed it came out on top in more tests than any other low-power SoC, and exhibited a 50 percent gain in energy efficiency from its debut results on MLPerf in April.

MLCommons said this round of MLPerfTM Inference v2.1 figures established new benchmarks with about 5,300 performance results submitted and 2,400 power measures, both up from the last set of results published.

Not surprisingly, Nvidia approves of MLPerf, saying companies participate in the tests because it is a valuable tool for vendors and customers evaluating AI platforms. Not everyone agrees it is the best way to measure machine learning performance, however.

Earlier this year, rival benchmark organization SPEC announced it had formed a committee to oversee the development of vendor-agnostic benchmarks for machine learning training and inference tasks. The organization said it intended to come up with benchmarks that will better represent industry practices than existing benchmarks such as MLPerf. ®

Send us news
Post a comment

AMD slaps together a silicon sandwich with MI300-series APUs, GPUs to challenge Nvidia’s AI empire

Chips boast 1.3x lead in AI, 1.8x in HPC over Nv's H100

HPE targets enterprises with Nvidia-powered platform for tuning AI

'We feel like enterprises are either going to become AI powered, or they're going to become obsolete'

Tech world forms AI Alliance to promote open and responsible AI

Everyone from Linux Foundation to NASA and Intel ... but some big names in AI are MIA

Creating a single AI-generated image needs as much power as charging your smartphone

PLUS: Microsoft to invest £2.5B in UK datacenters to power AI, and more

Mere minority of orgs put GenAI in production after year of hype

Folks are dipping their toes in without a full commitment

Don't be fooled: Google faked its Gemini AI voice demo

PLUS: The AI companies that will use AMD's latest GPUs, and more

Trust us, says EU, our AI Act will make AI trustworthy by banning the nasty ones

Big Tech plays the 'this might hurt innovation' card for rules that bar predictive policing, workplace emotion assessments

Nvidia’s China-market H20 chips hit another speed bump

Integration woes delay Nvidia's hopes of maintaining grip on Middle Kingdom

Dell APJ chief: Industry won't wait for Nvidia H100

Canalys mostly agrees, but thinks GPU giant still has a way to go

AWS unveils core-packed Graviton4 and beefier Trainium accelerators for AI

Also hedging its bets with a healthy dose of Nvidia chips too

Exposed Hugging Face API tokens offered full access to Meta's Llama 2

With more than 1,500 tokens exposed, research highlights importance of securing supply chains in AI and ML

Nvidia intros the 'SuperNIC' – it's like a SmartNIC, DPU or IPU, but more super

If you're doing AI but would rather not do InfiniBand, this NIC is for you