NVIDIA H100
Tensor Core GPU
EXCEPTIONAL PERFORMANCE,
SCALABILITY, AND SECURITY
FOR ALL DATA CENTERS

EXCEPTIONAL PERFORMANCE,
SCALABILITY, AND SECURITY
FOR ALL DATA CENTERS
| Form Factor | H100 SXM | H100 PCIe | H100 NVL² |
|---|---|---|---|
| FP64 | 34 teraFLOPS | 26 teraFLOPS | 68 teraFLOPs |
| FP64 Tensor Core | 67 teraFLOPS | 51 teraFLOPS | 134 teraFLOPs |
| FP32 | 67 teraFLOPS | 51 teraFLOPS | 134 teraFLOPs |
| TF32 Tensor Core | 989 teraFLOPS² | 756 teraFLOPS² | 1,979 teraFLOPs² |
| BFLOAT16 Tensor Core | 1,979 teraFLOPS² | 1,513 teraFLOPS² | 3,958 teraFLOPs² |
| FP16 Tensor Core | 1,979 teraFLOPS² | 1,513 teraFLOPS² | 3,958 teraFLOPs² |
| FP8 Tensor Core | 3,958 teraFLOPS² | 3,026 teraFLOPS² | 7,916 teraFLOPs² |
| INT8 Tensor Core | 3,958 TOPS² | 3,026 TOPS² | 7,916 TOPS² |
| GPU memory | 80GB | 80GB | 188GB |
| GPU memory bandwidth | 3.35TB/s | 2TB/s | 7.8TB/s³ |
| Decoders | 7 NVDEC 7 JPEG | 7 NVDEC 7 JPEG | 14 NVDEC 14 JPEG |
| Max thermal design power (TDP) | Up to 700W (configurable) | 300-350W (configurable) | 2x 350-400W (configurable) |
| Multi-Instance GPUs | Up to 7 MIGS @ 10GB each | Up to 7 MIGS @ 10GB each | Up to 14 MIGS @ 12GB each |
| Form factor | SXM | PCIe dual-slot air-cooled | 2x PCIe dual-slot air-cooled |
| Interconnect | NVLink: 900GB/s PCIe Gen5: 128GB/s | NVLink: 600GB/s PCIe Gen5: 128GB/s | NVLink: 600GB/s PCIe Gen5: 128GB/s |
| Server options |
NVIDIA HGX H100 Partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs NVIDIA DGX H100 with 8 GPUs |
Partner and NVIDIA-Certified Systems with 1–8 GPUs |
Partner and NVIDIA-Certified Systems with 2-4 pairs |
| NVIDIA AI Enterprise | Add-on | Included | Included |
Unlock an unprecedented
VDI user experience.
| GPU Memory | 4x 16GB GDDR6 with error-correcting code (ECC) |
| GPU Memory Bandwidth | 4x 200 GB/s |
| Max power consumption | 250W |
| Interconnect | PCI Express Gen 4 x16 |
| Form factor | Full height, full length (FHFL) dual slot |
| Thermal | Passive |
| vGPU Software Support | NVIDIA Virtual PC (vPC), NVIDIA Virtual Applications (vApps), NVIDIA RTX Virtual Workstation (vWS), NVIDIA Virtual Compute Server (vCS), and NVIDIA AI Enterprise |
| vGPU Profiles Supported | See the Virtual GPU Licensing Guide See the NVIDIA AI Enterprise Licensing Guide |
| NVENC | NVDEC | 4x | 8x (includes AV1 decode) |
| Secure and measured boot with hardware root of trust | Yes (optional) |
| NEBS Ready | Level 3 |
| Power Connector | 8-pin CPU |
AI-accelerated graphics
and visuals for mainstream
enterprise servers
| FP32 | 31.2 teraFLOPS |
| TF32 Tensor Core | 62.5 teraFLOPS | 125 teraFLOPS* |
| BFLOAT16 Tensor Core | 125 teraFLOPS | 250 teraFLOPS* |
| FP16 Tensor Core | 125 teraFLOPS | 250 teraFLOPS* |
| INT8 Tensor Core | 250 TOPS | 500 TOPS* |
| INT4 Tensor Core | 500 TOPS | 1,000 TOPS* |
| RT Core | 72 RT Cores |
| Encode/decode | 1 encoder 2 decoder (+AV1 decode) |
| GPU memory | 24GB GDDR6 |
| GPU memory bandwidth | 600GB/s |
| Interconnect | PCIe Gen4 64GB/s |
| Form factors | Single-slot, full-height, full-length (FHFL) |
| Max thermal design power (TDP) | 150W |
| vGPU software support | NVIDIA Virtual PC, NVIDIA Virtual Applications, NVIDIA RTX Virtual Workstation,NVIDIA Virtual Compute Server, NVIDIA AI Enterprise |
Entry-level GPUs that bring
NVIDIA AI to any server
| Peak FP32 | 4.5 TF |
| TF32 Tensor Core | 9 TF | 18 TF¹ |
| BFLOAT16 Tensor Core | 18 TF | 36 TF¹ |
| Peak FP16 Tensor Core | 18 TF | 36 TF¹ |
| Peak INT8 Tensor Core | 36 TOPS | 72 TOPS¹ |
| Peak INT4 Tensor Core | 72 TOPS | 144 TOPS¹ |
| RT Cores | 10 |
| Media engines | 1 video encoder 2 video decoders (includes AV1 decode) |
| GPU memory | 16GB GDDR6 |
| GPU memory bandwidth | 200GB/s |
| Interconnect | PCIe Gen4 x8 |
| Form factor | 1-slot, low-profile PCIe |
| Max thermal design power (TDP) | 40–60W (configurable) |
| Virtual GPU (vGPU) software support² | NVIDIA Virtual PC (vPC), NVIDIA Virtual Applications (vApps), NVIDIA RTX Virtual Workstation (vWS), NVIDIA AI Enterprise, NVIDIA Virtual Compute Server (vCS) |
Unparalleled AI and graphics
performance for the data center
| GPU Architecture | NVIDIA Ada Lovelace architecture |
|---|---|
| GPU Memory | 48GB GDDR6 with ECC |
| Memory Bandwidth | 864GB/s |
| Interconnect Interface | PCIe Gen4 x16: 64GB/s bidirectional |
| NVIDIA Ada Lovelace Architecture-Based CUDA® Cores | 18,176 |
| NVIDIA Third-Generation RT Cores | 142 |
| NVIDIA Fourth-Generation Tensor Cores | 568 |
| RT Core Performance TFLOPS | 212 |
| FP32 TFLOPS | 91.6 |
| TF32 Tensor Core TFLOPS | 183 I 366* |
| BFLOAT16 Tensor Core TFLOPS | 362.05 I 733* |
| FP16 Tensor Core | 362.05 I 733* |
| FP8 Tensor Core | 733 I 1,466* |
| Peak INT8 Tensor TOPS Peak INT4 Tensor TOPS | 733 I 1,466* 733 I 1,466* |
| Form Factor | 4.4" (H) x 10.5" (L), dual slot |
| Display Ports | 4x DisplayPort 1.4a |
| Max Power Consumption | 350W |
| Power Connector | 16-pin |
The breakthrough universal
accelerator for efficient video,
AI, and graphics
| FP32 | 30.3 teraFLOPs |
| TF32 Tensor Core | 120 teraFLOPS* |
| FP16 Tensor Core | 242 teraFLOPS* |
| BFLOAT16 Tensor Core | 242 teraFLOPS* |
| FP8 Tensor Core | 485 teraFLOPs* |
| INT8 Tensor Core | 485 TOPs* |
| GPU memory | 24GB |
| GPU memory bandwidth | 300GB/s |
| NVENC | NVDEC | JPEG decoders | 2 | 4 | 4 |
| Max thermal design power (TDP) | 72W |
| Form factor | 1-slot low-profile, PCIe |
| Interconnect | PCIe Gen4 x16 64GB/s |
| Server options | Partner and NVIDIA-Certified Systems with 1–8 GPUs |
World-proven enterprise AI
| GPUs | 8x NVIDIA H100 Tensor Core GPUs |
|---|---|
| GPU memory | 640GB total |
| Performance | 32 petaFLOPS FP8 |
| NVIDIA® NVSwitch™ | 4x |
| System power usage | 10.2kW max |
| CPU |
Dual Intel® Xeon® Platinum 8480C Processors 112 Cores total, 2.00 GHz(Base), 3.80 GHz (Max Boost) |
| System memory | 2TB |
| Networking |
4x OSFP ports serving 8x single-port NVIDIA ConnectX-7 VPI > Up to 400Gb/s InfiniBand/Ethernet 2x dual-port QSFP112 NVIDIA ConnectX-7 VPI > Up to 400Gb/s InfiniBand/Ethernet |
| Management Networking |
10Gb/s onboard NIC with RJ45 100Gb/s Ethernet NIC Host baseboard management controller (BMC) with RJ45 |
| Storage | OS: 2x 1.92TB NVMe M.2 |
| Internal storage | 8x 3.84TB NVMe U.2 |
| Software |
NVIDIA AI Enterprise – Optimized AI software NVIDIA Base Command – Orchestration, scheduling, and cluster management DGX OS / Ubuntu / Red Hat Enterprise Linux / Rocky – Operating System |
| Support | Comes with 3-year business-standard hardware and software support |
| System weight | 287.6lbs (130.45kgs) |
| Packaged system weight | 376lbs (170.45kgs) |
| System dimensions | Height: 14.0in (356mm) Width: 19.0in (482.2mm) Length: 35.3in (897.1mm) |
| Operating temperature range | 5–30°C (41–86°F) |