Build & Run AI/ML Models on the Most Powerful GPUs
NVIDIA H200 supercharges generative AI and HPC
High-performance LLM inference
H200 doubles inference performance compared to H100 when handling LLMs such as Llama2 70B. Get the highest throughput at the lowest TCO when deployed at scale for a massive user base.
Industry-leading generative AI training and fine-tuning
NVIDIA H200 GPUs feature the Transformer Engine with FP8 precision, which provides up to 5X faster training and 5.5X faster fine-tuning over A100 GPUs for large language models.
01
NVIDIA H100 features fourth-generation Tensor Cores and a Transformer Engine with FP8 precision that provides up to 4X faster training over the prior generation for GPT-3 (175B) models.
02
Built with 80 billion transistors using a cutting-edge TSMC 4N process custom tailored for NVIDIA’s accelerated compute needs, H100 features major advances to accelerate AI, HPC, memory bandwidth, interconnect, and communication at data center scale.