BACK TO INSIGHTS

Liquid vs. Air: Managing H100 Thermals

JULY 10, 2025 HARDWARE 6 MIN READ

The NVIDIA H100 SXM5 has a TDP of 700W. A standard 8-GPU chassis pulls over 6kW continuously. In a standard 42U rack, air cooling is no longer a viable engineering strategy. It is time to talk about plumbing.

The Physics of Air Failure

Air has a low specific heat capacity. To cool a dense rack of H100s (approx 40kW - 60kW per rack), you need hurricane-force airflow. This creates two problems:

Direct-to-Chip (DLC) Implementation

At Iteronix, our standard deployment for H100 clusters is Single-Phase Direct-to-Chip Liquid Cooling. We replace standard heatsinks with cold plates.

# Iteronix Thermal Spec (Rack A)
Coolant_Inlet_Temp: 32C
Flow_Rate: 1.5 LPM/GPU
PUE_Target: 1.08
Fan_Speed: 20% (Idle/Redundant)

The ROI of Liquid

While the CapEx of a liquid loop (CDU + Manifolds) is roughly 15% higher than air, the OpEx reduction is drastic. By removing the need for CRAC (Computer Room Air Conditioning) units to fight the GPUs, we see a 40% reduction in monthly electricity costs for the same compute output.

Deploying High Density?

We engineer the hydraulic loops for sovereign AI clusters.

Consult on Thermodynamics