The H200 features 141GB of HBM3e memory with 4.8 TB/s bandwidth, nearly double the capacity of the H100 with 1.4X more memory bandwidth. This enables processing of larger models and datasets without memory bottlenecks.
Superior AI Performance
Delivers 32 petaFLOPS of AI performance with 2X faster networking than the previous generation, making it ideal for the most demanding generative AI, large language models, and deep learning workloads.
Enhanced Compute Power
With up to 30% more CUDA cores and improved Tensor Core technology, the H200 provides a substantial performance boost over the H100, enabling faster training times and more efficient AI model development across various applications.
Use Cases
Large Model Inference
Run massive models with predictable latency. Optimize for throughput, batch size, and performance per watt.
Generative AI applications for text, image, and audio.
Scaling ML infrastructure as your customer base grows.