141GB
HBM3e Memory
4.8 TB/s
Memory Bandwidth
2600 TF
TensorRT FP8
700W
TGP
Technical Specifications
Enterprise-grade specifications for mission-critical AI workloads
Architecture & Performance
ArchitectureNVIDIA Hopper H200
Manufacturing ProcessTSMC 4N (4nm)
Memory141GB HBM3e
Memory Bandwidth4.8 TB/s
Memory Bus5120-bit
L2 Cache51MB
TensorRT FP82600 TF
Power & Form Factor
Form FactorSXM5 Module
Total Graphics Power700W
NVLink900 GB/s (4th Gen)
PCIePCIe 5.0 x16
Thermal SolutionPassive (Server)
Server SupportHGX H200 8-GPU
Precision SupportFP64, FP32, FP16, BF16, FP8
Enterprise Use Cases
Optimized for the most demanding AI and HPC workloads
Large Language Models
Train and deploy massive transformer models like GPT, LLaMA, and PaLM with 141GB memory per GPU enabling larger model sizes.
- • GPT-4 class model training
- • Multi-modal AI development
- • In-context learning at scale
Generative AI
Power next-generation generative AI applications with massive memory and bandwidth for real-time inference.
- • Real-time image generation
- • Video synthesis and editing
- • Code generation platforms
HPC & Scientific Computing
Accelerate computational fluid dynamics, molecular modeling, and weather simulation with enhanced FP64 performance.
- • Climate modeling
- • Molecular dynamics
- • Quantum simulations