Nvidia unveils a new GPU architecture designed for AI data centers

Nvidia’s H100 GPU is the first in its new family of Hopper processors that the company claims will underpin the world’s fastest supercomputer.

nvidia hopper gpu

While the rest of the computing industry struggles to get to one exaflop of computing, Nvidia is about to blow past everyone with an 18-exaflop supercomputer powered by a new GPU architecture.

The H100 GPU, has 80 billion transistors (the previous generation, Ampere, had 54 billion) with nearly 5TB/s of external connectivity and support for PCIe Gen5, as well as High Bandwidth Memory 3 (HBM3), enabling 3TB/s of memory bandwidth, the company says. It is the first in a new family of GPUs codenamed “Hopper,” after Admiral Grace Hopper, the computing pioneer who created COBOL and coined the term “computer bug.” It is due in the third quarter.

This GPU is meant to power data centers designed to handle heavy AI workloads, and Nvidia claims that 20 of them could sustain the equivalent of the entire world’s Internet traffic.

Hopper also comes with the second generation of Nvidia’s Secure Multi-Instance GPU (MIG) technology, allowing a single GPU to be partitioned to support security in multi-tenant uses. The key change with H100 is the MIGs are now fully isolated with I/O virtualization and independently secured with confidential computing capabilities each instance.

Researchers with smaller workloads were required to rent a full A100 CSP instance for isolation. With H100, they can use MIG to securely isolate a portion of a GPU, being assured that their data is secure.

To continue reading this article register now