The NVIDIA HGX H200 GPU was unveiled today, and it’s designed for generative AI and high performance computing workloads. Based on NVIDIA Hopper architecture, H200 is the first GPU to come equipped with HBM3e capable of delivering 141GB of memory at 4.8 terabytes per second.
NVIDIA H200 will be offered in NVIDIA HGX H200 server boards with four- and eight-way variants, both of which are compatible with the hardware as well as software of HGX H100 systems. It is also set to be available in the NVIDIA GH200 Grace Hopper Superchip with HBM3e, enabling it to be deployed in every type of data center, including on premises, cloud, hybrid-cloud as well as edge.
- NVIDIA Ampere Streaming Multiprocessors: The building blocks for the world’s fastest, most efficient GPU, the all-new Ampere SM brings 2X the FP32...
- 2nd Generation RT Cores: Experience 2X the throughput of 1st gen RT Cores, plus concurrent RT and shading for a whole new level of ray tracing...
- 3rd Generation Tensor Cores: Get up to 2X the throughput with structural sparsity and advanced AI algorithms such as DLSS. Now with support for up to...

To create intelligence with generative AI and HPC applications, vast amounts of data must be efficiently processed at high speed using large, fast GPU memory. With NVIDIA H200, the industry’s leading end-to-end AI supercomputing platform just got faster to solve some of the world’s most important challenges,” said Ian Buck, vice president of hyperscale and HPC at NVIDIA.
[Source]