Hopper

NVIDIA H200 - Inference-Optimized Hopper

NVIDIA H200 - Inference-Optimized Hopper

Complete specs, benchmarks, and analysis of the NVIDIA H200 - the HBM3e-equipped Hopper GPU that delivers 76% more memory and 43% more bandwidth than the H100 for inference workloads.