H100 NVL
NVIDIA · NVIDIA Hopper™ · Released 2022-03-27 · 16,896 (Per GPU) CUDA Cores · 94 VRAM · 350-400W (configurable) TDP

OVERVIEW
The NVIDIA H100 NVL variant is optimized for large language model inference, offering up to 5x performance improvement over NVIDIA A100 systems for LLMs up to 70 billion parameters. It features a PCIe form factor, NVLink bridge, and 188GB HBM3 memory for enhanced performance and scalability.
WORKLOAD SUITABILITY
The H100 NVL variant is best suited for large language model inference tasks, offering high performance and low latency for power-constrained data center environments.
- Large Language Model Inference
- AI Inference Acceleration
- Power-Constrained Data Centers
SPECIFICATIONS
WHAT THIS GPU IS GOOD AT
The H100 NVL excels at large-scale AI training and inference tasks, particularly in natural language processing and deep learning models. Its architecture is optimized for transformer models, offering significant performance improvements over previous generations. The GPU's high memory bandwidth and advanced tensor cores make it ideal for demanding computational workloads.
SERVER OPTIONS
The NVIDIA H100 NVL is available in high-performance server platforms such as NVIDIA's DGX systems and select OEM servers like Dell PowerEdge and HPE ProLiant. It is also offered in cloud instances, including AWS p5 and Azure ND series, providing scalable solutions for AI and HPC workloads.
POWER, THERMALS & NOISE
The H100 NVL has a high TDP, typically requiring advanced cooling solutions such as liquid cooling to maintain optimal performance. It is designed for datacenter environments where noise is less of a concern, but efficient thermal management is crucial to prevent throttling and ensure reliability.
COMPATIBILITY & SYSTEM FIT
This GPU is available in the SXM form factor, supporting NVLink for high-speed interconnects between multiple GPUs. It requires a compatible server platform with sufficient power delivery and cooling capabilities. The H100 NVL is designed for PCIe Gen 5.0 systems, ensuring maximum data throughput.
LIMITATIONS & KNOWN TRADE-OFFS
The H100 NVL's high power requirements and need for advanced cooling solutions can be a limitation for some deployments. Additionally, its premium pricing and availability constraints may pose challenges for smaller organizations. Users should also consider the infrastructure investment needed to fully leverage its capabilities.
PRICING
NOTES
Optimal for securely accelerating all workloads from enterprise to exascale.
"The H100 NVL variant features fourth-generation Tensor Cores and the Transformer Engine with FP8 precision."