NVIDIA Sets Six Records in AI Performance

This week NVIDIA announced that the company has broken a total of six performance records on a broad set of AI benchmarks. As a full suite, the benchmarks cover a variety of workloads and infrastructure scale – ranging from 16 GPUs on one node to up to 640 GPUs across 80 nodes.

Backed by Google, Intel, Baidu, NVIDIA and dozens more technology leaders, the new MLPerf benchmark suite measures a wide range of deep learning workloads. Aiming to serve as the industry’s first objective AI benchmark suite, it covers such areas as computer vision, language translation, personalized recommendations and reinforcement learning tasks.

The six categories include image classification, object instance segmentation, object detection, non-recurrent translation, recurrent translation and recommendation systems. NVIDIA did not submit results for the seventh category for reinforcement learning, which does not yet take advantage of GPU acceleration.

The new MLPerf benchmarks demonstrate the unmatched performance and versatility of NVIDIA’s Tensor Core GPUs,” said Ian Buck, vice president and general manager of Accelerated Computing at NVIDIA. “Exceptionally affordable and available in every geography from every cloud service provider and every computer maker, our Tensor Core GPUs are helping developers around the world advance AI at every stage of development.”

The software innovations and optimizations used to achieve NVIDIA’s industry-leading MLPerf performance are available free of charge in the company’s latest NGC deep learning containers.

Resource Links:

Latest Video

Industry Perspectives

AI is a game changer for industries today but achieving AI success contains two critical factors to consider — time to value and time to insights. Time to value is the metric that looks at the time it takes to realize the value of a product, solution or offering. Time to insight is a key measure for how long it takes to gain value from use of the product, solution or offering. [READ MORE…]

White Papers

XTREME-D’s next step in offering seamless HPC/DA/DL computing is the XTREME-Stargate gateway platform, or cluster portal. This on-prem device acts as a cluster “head node” and provides secure and fast access to bare-metal clusters that are configured using enhanced XTREME-DNA technology. Download the full paper to learn more about how XTREME-Stargate offers a “super head node” for HPC cloud clusters.