Skip to main content
Electrical 14 min read Mar 9, 2026

Data Center Power and Cooling: A Practical Engineering Guide

PUE fundamentals, ASHRAE thermal envelopes, cooling architectures, and efficiency strategies for modern facilities.

Data centers rank among the most energy-intensive building types in operation today. A single 10 MW facility can consume as much electricity as a small town, and roughly 30 to 40 percent of that energy goes toward cooling IT equipment rather than performing useful work. The metric that captures this overhead, Power Usage Effectiveness (PUE), has become the industry standard for measuring facility efficiency. But PUE alone does not tell engineers how to design or improve a facility.

This guide covers the engineering fundamentals behind data center power and cooling systems, from ASHRAE thermal guidelines and cooling technology selection through redundancy tiers, AI-driven density challenges, and carbon footprint estimation. Whether you are designing a new build or optimizing an existing facility, these principles apply across every scale from edge deployments to hyperscale campuses.

PUE: What It Measures and What Good Looks Like

PUE is the ratio of total facility power to IT equipment power. A facility drawing 10 MW total with 7 MW going to servers has a PUE of 1.43. The remaining 3 MW covers cooling, lighting, power distribution losses, and everything else that is not direct IT load. A perfect PUE of 1.0 would mean zero overhead, which is physically impossible since power conversion and distribution always introduce losses.

Industry averages have improved steadily over the past decade. The Uptime Institute's 2024 survey reported an average PUE of 1.58 across all surveyed data centers, down from roughly 2.0 ten years ago. Hyperscale operators such as Google report fleet-wide PUE around 1.10. Most enterprise facilities operate between 1.4 and 1.8. A PUE below 1.3 is considered good for a traditional raised-floor data center, while achieving below 1.2 typically requires free cooling or liquid cooling.

To measure PUE accurately, you need continuous metering at two points: total facility power at the utility entrance and IT load power at the output of the uninterruptible power supply (UPS). Spot readings taken on a mild spring afternoon will not reflect your August performance. Annual PUE calculated from 12 months of continuous data is the only honest comparison metric. Many operators report partial-year numbers that flatter their cooling efficiency.

PUE has well-known limitations. It does not account for IT equipment utilization, so a server farm running at 10 percent CPU utilization looks the same as one running at 80 percent. It can be manipulated by reclassifying loads. And it treats all kilowatt-hours equally regardless of grid carbon intensity. Despite these flaws, PUE remains the most widely used efficiency metric because it is simple to measure and straightforward to compare across facilities.

Formula:

PUE = Total Facility Power / IT Equipment Power

Example: 10 MW total facility, 7 MW IT load = PUE of 1.43. The 3 MW difference represents cooling, power distribution losses, lighting, and other overhead.

Electrical

Data Center Power & Cooling Load Calculator

Calculate electrical power requirements and cooling loads for data center spaces with PUE estimation, redundancy planning, and carbon footprint analysis.

Launch Calculator →

ASHRAE Thermal Guidelines for IT Equipment

ASHRAE Technical Committee 9.9 publishes thermal guidelines that define recommended and allowable temperature and humidity ranges for IT equipment. These guidelines have expanded over time as server hardware has become more tolerant of wider environmental conditions, enabling more efficient cooling strategies.

The current recommended envelope for Class A1 equipment is 64.4 to 80.6 degrees F dry bulb at the server inlet, with a dew point range of 41.9 to 59 degrees F and maximum relative humidity of 60 percent. This is where most facilities should target for normal operations. The allowable envelopes extend further: A1 allowable goes from 59 to 89.6 degrees F, A2 from 50 to 95 degrees F, A3 from 41 to 104 degrees F, and A4 from 41 to 113 degrees F.

The practical significance is straightforward. If your servers are rated for A2 or A3 conditions, you can run much warmer supply air temperatures, which dramatically increases the number of hours per year when outside air or evaporative cooling can handle the load without mechanical refrigeration. Raising the supply air setpoint from 68 degrees F to 80 degrees F can reduce cooling energy by 30 to 40 percent in temperate climates.

However, running at the upper end of the allowable range does increase component failure rates. Published data from hyperscale operators show that server failure rates increase roughly 2 to 4 percent for every 18 degrees F above the recommended range. The practical sweet spot for most operators is supply air in the 75 to 80 degrees F range, capturing most of the free cooling benefit without significantly increasing hardware replacement costs.

Tip:

Check the ASHRAE class rating on your server hardware before designing for elevated inlet temperatures. Most enterprise servers from major OEMs are rated A2 or higher, but storage arrays and network equipment may have tighter thermal requirements that limit your overall setpoint.

Cooling Technologies: Choosing the Right Architecture

Direct expansion (DX) systems are the simplest cooling option. Packaged units with compressors, condensers, and evaporator coils sit in or adjacent to the data hall. DX works well for small and medium facilities under 500 kW of IT load. These systems are quick to install, but they top out in efficiency around a PUE of 1.5 to 1.7 because the compressors run continuously with no easy path to use cool outside air.

Chilled water systems scale better for larger facilities. Central chillers produce cold water at 42 to 50 degrees F, which is piped to computer room air handlers (CRAHs) on the data center floor. The advantages include centralized maintenance, modular capacity expansion, and compatibility with waterside economizers. A well-designed chilled water plant with variable-speed chillers, optimized condenser water temperature, and waterside economizer can achieve PUE below 1.3.

Evaporative cooling uses the latent heat of water evaporation to reject heat without mechanical refrigeration. Direct evaporative coolers add moisture to the supply air, which works well in arid climates but adds humidity in already-humid regions. Indirect evaporative systems use a heat exchanger to cool supply air without adding moisture. Hyperscale operators have pushed indirect evaporative cooling to achieve PUE values of 1.06 to 1.12 in favorable climates. The tradeoff is water consumption, typically 1.5 to 3 gallons per kWh of cooling.

Liquid cooling is increasingly necessary for high-density AI and GPU workloads. Direct-to-chip cold plates circulate coolant directly over processors, rejecting heat at 104 to 140 degrees F water temperatures. This elevated return temperature enables extremely efficient heat rejection using dry coolers without chillers for most of the year. Immersion cooling submerges entire servers in dielectric fluid and can handle power densities above 100 kW per rack, compared to 10 to 20 kW per rack for traditional air cooling.

Evaporative cooling water consumption is a growing concern in water-stressed regions. A 5 MW facility using direct evaporative cooling may consume 3 to 5 million gallons per year. Dry cooler and liquid cooling approaches use zero water for heat rejection.

Electrical

Data Center Power & Cooling Load Calculator

Calculate electrical power requirements and cooling loads for data center spaces with PUE estimation, redundancy planning, and carbon footprint analysis.

Launch Calculator →

Free Cooling: Hours by Climate Zone

Free cooling, also called economization, uses outdoor conditions to cool the data center without running mechanical refrigeration. The economics are simple: every hour the chillers can stay off saves roughly 30 to 40 percent of cooling energy for that period. The key design question is how many hours per year your climate allows it.

For airside economization at an 80 degrees F supply setpoint, approximate annual free cooling hours by U.S. climate zone are as follows. Climate Zone 1 (Miami): about 200 hours. Zone 3 (Atlanta): about 3,500 hours. Zone 4 (New York): about 5,000 hours. Zone 5 (Chicago): about 5,800 hours. Zone 6 (Minneapolis): about 6,500 hours. Zone 7 and 8 (Duluth, Fairbanks): about 7,500 or more hours. A facility in Minneapolis can run on free cooling roughly 75 percent of the year. A facility in Miami gets almost none.

Waterside economization uses a plate-and-frame heat exchanger to transfer heat from the chilled water loop to the condenser water loop when the wet-bulb temperature is low enough. This approach provides partial free cooling at higher outdoor temperatures than airside economization, extending the useful hours. Many facilities in Zone 4 to 5 climates achieve 3,000 to 4,000 hours of full economizer operation and another 2,000 hours of partial economization where the heat exchanger handles part of the load and the chillers handle the remainder.

The decision between airside and waterside economization involves tradeoffs beyond hours of operation. Airside economizers introduce outdoor air contaminants (particulates, humidity, gaseous pollutants) that can damage IT equipment. Filtration and humidity control add cost and maintenance. Waterside economizers keep the data hall sealed and controlled but require more mechanical infrastructure and typically cannot achieve PUE values as low as well-executed airside economization in cool climates.

Tip:

When evaluating a new site location, pull TMY3 weather data for the area and calculate annual economizer hours at your planned supply temperature. The difference between two candidate sites can be thousands of hours per year, translating to hundreds of thousands of dollars in annual cooling costs.

Redundancy Levels and Their Cost Impact

Data center redundancy is described by the Uptime Institute's tier classification system, which defines four levels of infrastructure redundancy. Understanding the cost implications of each tier is critical for right-sizing the investment to actual business requirements.

Tier I (Basic): No redundancy, single path for power and cooling. Expected uptime of 99.671 percent, or about 28.8 hours of downtime per year. Construction cost roughly $7 to $10 per watt of IT load. Appropriate for development and test environments where brief outages are tolerable.

Tier II (Redundant Components): N+1 redundancy on major components such as an extra UPS module or cooling unit. Single distribution path. Expected uptime of 99.741 percent, about 22 hours of downtime per year. Construction cost roughly $10 to $14 per watt. Suitable for small business applications with moderate availability needs.

Tier III (Concurrently Maintainable): Dual power and cooling distribution paths, only one active at a time. Any component can be removed for maintenance without affecting IT load. Expected uptime of 99.982 percent, about 1.6 hours of downtime per year. Construction cost roughly $15 to $22 per watt. This is the most common tier for enterprise and colocation facilities.

Tier IV (Fault Tolerant): Dual active distribution paths. The facility can sustain any single infrastructure failure without affecting IT load. Expected uptime of 99.995 percent, about 0.4 hours of downtime per year. Construction cost roughly $22 to $30 or more per watt. Reserved for mission-critical applications like financial trading and healthcare systems.

Warning:

The cost difference between Tier II and Tier IV can be 2 to 3 times, and operational complexity scales accordingly. Most facilities do not need Tier IV. Define your actual uptime requirements with the business before committing to a tier level.

AI and GPU Rack Density Challenges

Traditional server racks consume 5 to 8 kW each. Virtualization and higher-density servers pushed typical enterprise racks to 10 to 15 kW over the past decade. AI and machine learning workloads have pushed far beyond those levels. A rack of NVIDIA H100 GPUs can draw 40 to 70 kW, and next-generation GPU platforms are projected to push individual racks above 100 kW.

This density shift has fundamental implications for cooling architecture. Air cooling works reliably up to about 25 to 30 kW per rack with proper hot/cold aisle containment. Beyond that threshold, the volume of air required becomes impractical. A 60 kW rack needs roughly 10,000 CFM of supply air at a 20 degrees F temperature differential. At that airflow, noise levels become extreme and the fan energy alone begins to erode PUE gains.

Liquid cooling becomes the only viable path for racks above 30 kW. Direct-to-chip cold plates handle the GPU and CPU heat loads while fans still cool memory, storage, and other board-level components. Full immersion cooling eliminates fans entirely and can handle the highest densities, but it requires significant changes to server design, maintenance procedures, and facility plumbing.

Power distribution also changes at high densities. A 60 kW rack at 208V three-phase draws about 166 amps, which exceeds the capacity of standard C13/C14 power connections. High-density deployments typically use busway distribution or high-amperage whip connections. Transformer and UPS sizing must account for concentrated per-rack loads and the possibility of rapid load changes as GPU training jobs start and stop.

When planning for AI workloads, design the cooling infrastructure for the full rack power draw, not the average. GPU training jobs can swing from idle to full load in seconds, and the cooling system must respond without letting inlet temperatures exceed ASHRAE limits.

Electrical

Data Center Power & Cooling Load Calculator

Calculate electrical power requirements and cooling loads for data center spaces with PUE estimation, redundancy planning, and carbon footprint analysis.

Launch Calculator →

Estimating Your Facility's Carbon Footprint

A data center's carbon footprint is driven by two factors: total energy consumption and the carbon intensity of the electricity source. The EPA's eGRID database provides regional emission factors that make this calculation straightforward. A facility in the RFCW region (Ohio, Indiana, West Virginia) generates roughly 1.3 lbs of CO2 per kWh due to heavy coal and natural gas generation. A facility in the NWPP region (Pacific Northwest) generates about 0.5 lbs CO2 per kWh thanks to hydroelectric power.

Moving the same workload from Ohio to Oregon cuts its carbon footprint by more than 60 percent without changing a single piece of equipment. This is why site selection has become a sustainability decision as much as a real estate and connectivity decision. Regions with clean grids and cool climates (Pacific Northwest, Nordics, parts of Canada) offer both lower carbon intensity and more free cooling hours.

For operators who cannot relocate, power purchase agreements (PPAs) for off-site wind or solar generation are the primary mechanism for reducing reported carbon intensity. Renewable energy certificates (RECs) provide another option, though their additionality (whether they actually cause new renewable capacity to be built) is debated. On-site solar can offset daytime load but rarely covers more than 10 to 20 percent of a data center's total consumption due to the mismatch between roof area and power density.

Scope 3 emissions from embodied carbon in construction materials and IT hardware manufacturing are gaining attention. The carbon footprint of manufacturing a server can equal 1 to 2 years of its operational electricity consumption. Extending server refresh cycles from 3 to 5 years reduces embodied carbon significantly, though it must be balanced against the efficiency gains of newer hardware.

Tip:

Use the EPA's eGRID subregion for your facility location, not the state average. Emission factors can vary significantly within a state depending on which grid subregion serves your area.

Frequently Asked Questions

Industry average is about 1.58. A PUE below 1.4 is considered good for a traditional enterprise facility. Below 1.2 is excellent and typically requires free cooling or liquid cooling. Hyperscale operators achieve 1.06 to 1.12. For a new build, targeting 1.2 to 1.3 is realistic with chilled water, economizer, and containment.
Liquid cooling becomes necessary above roughly 25 to 30 kW per rack, where air cooling volume becomes impractical. For AI and GPU racks drawing 40 to 100 kW or more, liquid cooling is the only viable option. Direct-to-chip liquid cooling adds $200 to $500 per kW of IT capacity in infrastructure cost but reduces cooling energy costs by 30 to 50 percent at high densities.
Each successive ASHRAE class allows wider temperature and humidity ranges at the server inlet. A1 recommended range tops out at 80.6 degrees F. A2 allowable extends to 95 degrees F, A3 to 104 degrees F. Wider ranges mean more free cooling hours per year. If your servers are rated A2 or higher, you can design for elevated supply temperatures and capture significantly more economizer hours.
Direct evaporative cooling typically consumes 1.5 to 3 gallons of water per kWh of heat rejected. A 5 MW data center using evaporative cooling might consume 3 to 5 million gallons per year. In water-scarce regions, this is a real constraint. Dry coolers and liquid-to-air systems use zero water for heat rejection.
Tier IV typically costs 40 to 60 percent more than Tier III for power and cooling infrastructure. Tier III provides concurrent maintainability with dual distribution paths (one active). Tier IV provides fault tolerance with both paths simultaneously active. Most enterprise workloads are well-served by Tier III.
Multiply your annual electricity consumption in kWh by the EPA eGRID emission factor for your grid subregion. A 5 MW facility at PUE 1.4 consuming about 61 million kWh per year in a coal-heavy region at 1.3 lbs CO2 per kWh produces roughly 40,000 tons of CO2 annually. Facilities in hydro-powered regions produce 60 percent less.
Hot and cold aisle containment is almost always the best first step. Preventing hot exhaust air from mixing with cold supply air typically improves PUE by 0.1 to 0.3 at a cost of $50 to $200 per rack. Payback is usually under one year. After containment, raising supply air temperature and enabling economizer operation are the next highest-impact measures.
Disclaimer: This guide provides general engineering information about data center power and cooling design. Specific facility designs must account for local climate, utility infrastructure, code requirements, and business continuity needs. Consult qualified mechanical and electrical engineers for facility-specific design work.

Calculators Referenced in This Guide

Electrical Live

Panel Load Study

Do you actually need a panel upgrade? Walk your breaker panel with NEC Article 220 demand factors. See connected load vs. calculated demand and test whether an EV charger, heat pump, or hot tub fits.

Electrical Live

Data Center Power & Cooling Load Calculator

Calculate electrical power requirements and cooling loads for data center spaces with PUE estimation, redundancy planning, and carbon footprint analysis.

Related Guides

Electrical 13 min

Battery Energy Storage Systems: Sizing, Chemistry, and Financial Analysis

BESS application modes, LFP vs NMC vs flow battery selection, degradation modeling, round-trip efficiency, IRA ITC incentives, LCOS calculation, and system sizing methodology.

Electrical 14 min

Transformer Loss Evaluation: Total Owning Cost and Efficiency Analysis

IEEE C57.120 A/B factor method, no-load vs load loss evaluation, DOE 2016 efficiency standards, K-factor harmonics impact, loading growth projections, and total owning cost comparison methodology.