H100 NVL

NVIDIA

NVIDIA · NVIDIA Hopper™ · Released 2022-03-27 · 16,896 (Per GPU) CUDA Cores · 94 VRAM · 350-400W (configurable) TDP

H100 NVL

OVERVIEW

The NVIDIA H100 NVL variant is optimized for large language model inference, offering up to 5x performance improvement over NVIDIA A100 systems for LLMs up to 70 billion parameters. It features a PCIe form factor, NVLink bridge, and 188GB HBM3 memory for enhanced performance and scalability.

WORKLOAD SUITABILITY

The H100 NVL variant is best suited for large language model inference tasks, offering high performance and low latency for power-constrained data center environments.

  • Large Language Model Inference
  • AI Inference Acceleration
  • Power-Constrained Data Centers

SPECIFICATIONS

CUDA Cores16,896 (Per GPU)
memory gb94
tgp watts350-400W (configurable)
cuda cores-
architectureNVIDIA Hopper™
precision compute
FP83,341 teraFLOPs
FP161,671 teraFLOPs
FP3260 teraFLOPs
FP6430 teraFLOPs
INT83,341 TOPS
TF32835 teraFLOPs
BFLOAT161,671 teraFLOPs
memory bandwidth gbps3.9

WHAT THIS GPU IS GOOD AT

The H100 NVL excels at large-scale AI training and inference tasks, particularly in natural language processing and deep learning models. Its architecture is optimized for transformer models, offering significant performance improvements over previous generations. The GPU's high memory bandwidth and advanced tensor cores make it ideal for demanding computational workloads.

SERVER OPTIONS

The NVIDIA H100 NVL is available in high-performance server platforms such as NVIDIA's DGX systems and select OEM servers like Dell PowerEdge and HPE ProLiant. It is also offered in cloud instances, including AWS p5 and Azure ND series, providing scalable solutions for AI and HPC workloads.

POWER, THERMALS & NOISE

The H100 NVL has a high TDP, typically requiring advanced cooling solutions such as liquid cooling to maintain optimal performance. It is designed for datacenter environments where noise is less of a concern, but efficient thermal management is crucial to prevent throttling and ensure reliability.

COMPATIBILITY & SYSTEM FIT

This GPU is available in the SXM form factor, supporting NVLink for high-speed interconnects between multiple GPUs. It requires a compatible server platform with sufficient power delivery and cooling capabilities. The H100 NVL is designed for PCIe Gen 5.0 systems, ensuring maximum data throughput.

LIMITATIONS & KNOWN TRADE-OFFS

The H100 NVL's high power requirements and need for advanced cooling solutions can be a limitation for some deployments. Additionally, its premium pricing and availability constraints may pose challenges for smaller organizations. Users should also consider the infrastructure investment needed to fully leverage its capabilities.

PRICING

Vendor
Price
Unit
Currency
Date Added
Crusoe Cloud
View listing →
$0.00
hour
USD
Feb 2, 2026
$3.07
hour
USD
Feb 2, 2026

NOTES

Optimal for securely accelerating all workloads from enterprise to exascale.

"The H100 NVL variant features fourth-generation Tensor Cores and the Transformer Engine with FP8 precision."