NVIDIA H100
Tensor Core GPU
 
                        EXCEPTIONAL PERFORMANCE, 
                        SCALABILITY, AND SECURITY 
                        FOR ALL DATA CENTERS
                    

                        EXCEPTIONAL PERFORMANCE, 
                        SCALABILITY, AND SECURITY 
                        FOR ALL DATA CENTERS
                    
| Form Factor | H100 SXM | H100 PCIe | H100 NVL² | 
|---|---|---|---|
| FP64 | 34 teraFLOPS | 26 teraFLOPS | 68 teraFLOPs | 
| FP64 Tensor Core | 67 teraFLOPS | 51 teraFLOPS | 134 teraFLOPs | 
| FP32 | 67 teraFLOPS | 51 teraFLOPS | 134 teraFLOPs | 
| TF32 Tensor Core | 989 teraFLOPS² | 756 teraFLOPS² | 1,979 teraFLOPs² | 
|  BFLOAT16 Tensor Core  |  1,979 teraFLOPS² | 1,513 teraFLOPS² | 3,958 teraFLOPs² | 
| FP16 Tensor Core | 1,979 teraFLOPS² | 1,513 teraFLOPS² | 3,958 teraFLOPs² | 
| FP8 Tensor Core | 3,958 teraFLOPS² | 3,026 teraFLOPS² | 7,916 teraFLOPs² | 
| INT8 Tensor Core | 3,958 TOPS² | 3,026 TOPS² | 7,916 TOPS² | 
| GPU memory | 80GB | 80GB | 188GB | 
|  GPU memory bandwidth  |  3.35TB/s | 2TB/s | 7.8TB/s³ | 
| Decoders |  7 NVDEC 7 JPEG  |   7 NVDEC 7 JPEG  |   14 NVDEC 14 JPEG  | 
|  Max thermal design power (TDP)  |   Up to 700W  (configurable)  |   300-350W  (configurable)  |   2x 350-400W (configurable)  | 
|  Multi-Instance GPUs  |   Up to 7 MIGS @ 10GB each  |   Up to 7 MIGS @ 10GB each  |   Up to 14 MIGS @ 12GB each  | 
| Form factor | SXM |  PCIe  dual-slot air-cooled  |   2x PCIe dual-slot air-cooled  | 
| Interconnect |  NVLink: 900GB/s  PCIe Gen5: 128GB/s  |   NVLink: 600GB/s  PCIe Gen5: 128GB/s  |   NVLink: 600GB/s  PCIe Gen5: 128GB/s  | 
| Server options |  
                                NVIDIA HGX H100 Partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs NVIDIA DGX H100 with 8 GPUs  |   
                                Partner and NVIDIA-Certified Systems with 1–8 GPUs  |  
                                Partner and NVIDIA-Certified Systems with 2-4 pairs  | 
| NVIDIA AI Enterprise | Add-on | Included | Included | 
Unlock an unprecedented 
                                VDI user experience.
| GPU Memory | 4x 16GB GDDR6 with error-correcting code (ECC) | 
| GPU Memory Bandwidth | 4x 200 GB/s | 
| Max power consumption | 250W | 
| Interconnect | PCI Express Gen 4 x16 | 
| Form factor | Full height, full length (FHFL) dual slot | 
| Thermal | Passive | 
| vGPU Software Support | NVIDIA Virtual PC (vPC), NVIDIA Virtual Applications (vApps), NVIDIA RTX Virtual Workstation (vWS), NVIDIA Virtual Compute Server (vCS), and NVIDIA AI Enterprise | 
| vGPU Profiles Supported | See the Virtual GPU Licensing Guide  See the NVIDIA AI Enterprise Licensing Guide  | 
| NVENC | NVDEC | 4x | 8x (includes AV1 decode) | 
| Secure and measured boot with hardware root of trust | Yes  (optional)  | 
| NEBS Ready | Level 3 | 
| Power Connector | 8-pin CPU | 
                                AI-accelerated graphics 
                                and visuals for mainstream 
                                enterprise servers
                            
| FP32 | 31.2 teraFLOPS | 
| TF32 Tensor Core | 62.5 teraFLOPS | 125 teraFLOPS* | 
| BFLOAT16 Tensor Core | 125 teraFLOPS | 250 teraFLOPS* | 
| FP16 Tensor Core | 125 teraFLOPS | 250 teraFLOPS* | 
| INT8 Tensor Core | 250 TOPS | 500 TOPS* | 
| INT4 Tensor Core | 500 TOPS | 1,000 TOPS* | 
| RT Core | 72 RT Cores | 
| Encode/decode | 1 encoder 2 decoder (+AV1 decode)  | 
| GPU memory | 24GB GDDR6 | 
| GPU memory bandwidth | 600GB/s | 
| Interconnect | PCIe Gen4 64GB/s | 
| Form factors | Single-slot, full-height, full-length (FHFL) | 
| Max thermal design power (TDP)  |  150W | 
| vGPU software support | NVIDIA Virtual PC, NVIDIA Virtual Applications,  NVIDIA RTX Virtual Workstation,NVIDIA Virtual Compute Server, NVIDIA AI Enterprise  | 
Entry-level GPUs that bring 
                                NVIDIA AI to any server
| Peak FP32 | 4.5 TF | 
| TF32 Tensor Core | 9 TF | 18 TF¹ | 
| BFLOAT16 Tensor Core | 18 TF | 36 TF¹ | 
| Peak FP16 Tensor Core | 18 TF | 36 TF¹ | 
| Peak INT8 Tensor Core | 36 TOPS | 72 TOPS¹ | 
| Peak INT4 Tensor Core | 72 TOPS | 144 TOPS¹ | 
| RT Cores | 10 | 
| Media engines | 1 video encoder 2 video decoders (includes AV1 decode)  | 
| GPU memory | 16GB GDDR6 | 
| GPU memory bandwidth | 200GB/s | 
| Interconnect | PCIe Gen4 x8 | 
| Form factor | 1-slot, low-profile PCIe | 
| Max thermal design power (TDP)  |  40–60W (configurable) | 
| Virtual GPU (vGPU) software support²  |  NVIDIA Virtual PC (vPC), NVIDIA Virtual Applications (vApps), NVIDIA RTX Virtual Workstation (vWS), NVIDIA AI Enterprise, NVIDIA Virtual Compute Server (vCS) | 
Unparalleled AI and graphics 
                                performance for the data center
| GPU Architecture | NVIDIA Ada Lovelace architecture | 
|---|---|
| GPU Memory | 48GB GDDR6 with ECC | 
| Memory Bandwidth | 864GB/s | 
| Interconnect Interface | PCIe Gen4 x16: 64GB/s bidirectional | 
| NVIDIA Ada Lovelace Architecture-Based CUDA® Cores  |  18,176 | 
| NVIDIA Third-Generation RT Cores  |  142 | 
| NVIDIA Fourth-Generation Tensor Cores  |  568 | 
| RT Core Performance TFLOPS  |  212 | 
| FP32 TFLOPS | 91.6 | 
| TF32 Tensor Core TFLOPS  |  183 I 366* | 
| BFLOAT16 Tensor Core TFLOPS  |  362.05 I 733* | 
| FP16 Tensor Core | 362.05 I 733* | 
| FP8 Tensor Core | 733 I 1,466* | 
| Peak INT8 Tensor TOPS Peak INT4 Tensor TOPS  |  733 I 1,466* 733 I 1,466*  | 
| Form Factor | 4.4" (H) x 10.5" (L), dual slot | 
| Display Ports | 4x DisplayPort 1.4a | 
| Max Power Consumption | 350W | 
| Power Connector | 16-pin | 
                                The breakthrough universal 
                                accelerator for efficient video, 
                                AI, and graphics
                            
| FP32 | 30.3 teraFLOPs | 
| TF32 Tensor Core | 120 teraFLOPS* | 
| FP16 Tensor Core | 242 teraFLOPS* | 
| BFLOAT16 Tensor Core | 242 teraFLOPS* | 
| FP8 Tensor Core | 485 teraFLOPs* | 
| INT8 Tensor Core | 485 TOPs* | 
| GPU memory | 24GB | 
| GPU memory bandwidth | 300GB/s | 
| NVENC | NVDEC | JPEG  decoders  |  2 | 4 | 4 | 
| Max thermal design power (TDP)  |  72W | 
| Form factor | 1-slot low-profile, PCIe | 
| Interconnect | PCIe Gen4 x16 64GB/s | 
| Server options | Partner and NVIDIA-Certified Systems with 1–8 GPUs | 
World-proven enterprise AI
| GPUs | 8x NVIDIA H100 Tensor Core GPUs | 
|---|---|
| GPU memory | 640GB total | 
| Performance | 32 petaFLOPS FP8 | 
| NVIDIA® NVSwitch™ | 4x | 
| System power usage | 10.2kW max | 
| CPU | 
                                Dual Intel® Xeon® Platinum 8480C  Processors 112 Cores total, 2.00 GHz(Base), 3.80 GHz (Max Boost)  | 
| System memory | 2TB | 
| Networking | 
                                4x OSFP ports serving 8x single-port NVIDIA ConnectX-7 VPI > Up to 400Gb/s InfiniBand/Ethernet 2x dual-port QSFP112 NVIDIA ConnectX-7 VPI > Up to 400Gb/s InfiniBand/Ethernet  | 
| Management Networking  |  
                                10Gb/s onboard NIC with RJ45 100Gb/s Ethernet NIC  Host baseboard management controller (BMC) with RJ45  | 
| Storage | OS: 2x 1.92TB NVMe M.2 | 
| Internal storage | 8x 3.84TB NVMe U.2 | 
| Software | 
                                NVIDIA AI Enterprise – Optimized AI software NVIDIA Base Command – Orchestration, scheduling, and cluster management DGX OS / Ubuntu / Red Hat Enterprise Linux / Rocky – Operating System  | 
| Support |  Comes with 3-year business-standard hardware and software support  | 
| System weight | 287.6lbs (130.45kgs) | 
| Packaged system weight | 376lbs (170.45kgs) | 
| System dimensions | Height: 14.0in (356mm) Width: 19.0in (482.2mm) Length: 35.3in (897.1mm)  | 
| Operating temperature range | 5–30°C (41–86°F) |