Skip to content
Longterm Wiki
Back

NVIDIA H100 GPU - Data Center Accelerator (Hopper Architecture)

web

The H100 GPU is the dominant hardware platform for frontier AI model training and deployment; relevant to AI governance discussions around compute access, hardware concentration, and the pace of AI capability scaling.

Metadata

Importance: 42/100tool pagereference

Summary

Official product page for NVIDIA's H100 GPU, built on the Hopper architecture, which represents a major leap in AI training and inference compute. The H100 delivers up to 4X faster AI training over the prior A100 generation and 30X faster inference for large language models, with features like a dedicated Transformer Engine and fourth-generation NVLink interconnect.

Key Points

  • H100 uses Hopper architecture with fourth-generation Tensor Cores and a dedicated Transformer Engine supporting FP8 precision for LLM training
  • Delivers up to 4X faster GPT-3 training and 30X faster LLM inference compared to the previous A100 generation
  • Supports trillion-parameter model training, enabling exascale AI workloads previously out of reach for most researchers
  • Features 900 GB/s NVLink interconnect and NDR InfiniBand networking for large-scale GPU cluster deployment
  • Represents a key component of the AI compute stack central to discussions about AI development pace and hardware concentration

Cited by 1 page

PageTypeQuality
AI Knowledge MonopolyRisk50.0

Cached Content Preview

HTTP 200Fetched Mar 15, 202615 KB
H100 GPU | NVIDIA 

 
 

 

 

 
 
 
 
 
 
 
 
 
 
 
 

 

 

 
 This site requires Javascript in order to view all its content. Please enable Javascript in order to access all the functionality of this web site. Here are the instructions how to enable JavaScript in your web browser. 
 
 
 
 

 
 

 
 
 

 
 
 

 
 
 

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

 
 
 

 

 

 
 

 
 
 

 
 
 

 
 

 

 
 

 
 
 
 
 
 
 
 
 
 NVIDIA H100 GPU
 
 

 
 
 
 

 
 
 
 
 
 
 
 
 Extraordinary performance, scalability, and security for every data center. 

 
 
 

 

 
 
 
 
 
 
 

 View Datasheet 
 
 
 

 

 
 

 
 
 
 

 
 
 

 

 

 
 

 
 
 
 

 
 
 

 

 
 
 
 
 
 
 
 
 
 Introduction 
 

 
 
 
 Workloads 
 

 
 
 
 H100 NVL 
 

 
 
 
 NVIDIA Enterprise 
 

 
 
 
 Specifications 
 

 
 
 
 
 
 
 
 Introduction 
 

 
 
 Workloads 
 

 
 
 H100 NVL 
 

 
 
 NVIDIA Enterprise 
 

 
 
 Specifications 
 

 
 
 
 
 
 
 
 
 Introduction 
 

 
 
 Workloads 
 

 
 
 H100 NVL 
 

 
 
 NVIDIA Enterprise 
 

 
 
 Specifications 
 

 
 
 
 
 
 
 
 
 

 
 
 

 
 

 

 
 

 
 
 
 
 
 
 
 
 
 An Order-of-Magnitude Leap for Accelerated Computing
 
 

 
 
 
 

 
 
 
 
 
 
 
 
 The NVIDIA H100 GPU delivers exceptional performance, scalability, and security for every workload. H100 uses breakthrough innovations based on the NVIDIA Hopper™ architecture to deliver industry-leading conversational AI, speeding up large language models (LLMs) by 30X. H100 also includes a dedicated Transformer Engine to solve trillion-parameter language models. 

 
 
 

 

 
 
 
 
 
 
 

 Read NVIDIA H100 Datasheet 
 
 
 

 
 
 
 
 
 
 

 NVIDIA H100 NVL Product Brief 
 
 
 

 

 
 

 
 
 
 

 
 
 

 

 
 

 
 
 

 
 

 

 
 

 
 
 
 
 
 
 
 
 
 Securely Accelerate Workloads From Enterprise to Exascale
 
 

 
 
 
 

 
 
 

 
 
 
 

 
 
 
 
 
 
 
 

 
 
 

 
 

 

 
 

 
 
 
 
 
 
 
 
 
 Up to 4X Higher AI Training on GPT-3
 
 

 
 
 
 

 
 
 

 
 
 
 
 
 
 
 
 

 
 
 
 
 

 
 
 
 
 
 
 
 
 Projected performance subject to change. GPT-3 175B training A100 cluster: HDR IB network, H100 cluster: NDR IB network | Mixture of Experts (MoE) Training Transformer Switch-XXL variant with 395B parameters on 1T token dataset, A100 cluster: HDR IB network, H100 cluster: NDR IB network with NVLink Switch System where indicated. 

 
 
 

 

 

 
 

 
 
 
 

 
 
 

 

 
 

 
 
 

 
 

 

 
 

 
 
 
 
 
 
 
 
 
 Transformational AI Training
 
 

 
 
 
 

 
 
 
 
 
 
 
 
 H100 features fourth-generation Tensor Cores and a Transformer Engine with FP8 precision that provides up to 4X faster training over the prior generation for GPT-3 (175B) models. The combination of fourth-generation NVLink, which offers 900 gigabytes per second (GB/s) of GPU-to-GPU interconnect; NDR Quantum-2 InfiniBand networking, which accelerates communication by every GPU across nodes; PCIe Gen5; and NVIDIA Magnum IO™ software delivers efficient scalability from small enterprise systems to massive, unified GPU clusters.

 
 Deploying H100 GPUs at dat

... (truncated, 15 KB total)
Resource ID: 0907d57e1be07428 | Stable ID: NTc0MzRlMG