Monday, September 22, 2025

"All of the Above" Helps AI Data Centers Reduce Energy Demand

Just as electrical utilities use rate differentials to shift consumer workloads to off-peak hours, so data centers supporting artificial intelligence jobs can use workload shaping to shift jobs to off-peak hours. 


"Demand-side" management options are just as important as “supply side” increases in energy infrastructure. These demand management measures include:

  • optimizing hardware

  • improving software and workload management

  • enhancing physical infrastructure.


By identifying which AI tasks are time-sensitive (such as real-time inference for a search engine) versus which are not (training a new large language model), data centers can dynamically shift computational loads. 


Non-critical tasks can be paused or slowed down during peak grid demand hours and resumed when electricity is cheaper, cleaner, or more abundant. 


In a related way, workloads can be scheduled to run when the local grid's energy mix is dominated by renewable sources like solar or wind, which can reduce overall consumption.


The AI models themselves can be designed to be more efficient.


Techniques such as quantization and pruning reduce a model's size and the number of calculations required without significantly compromising accuracy. For example, by converting a model's parameters from 32-bit to 8-bit, its energy needs can be drastically reduced.


For certain tasks, an AI model can be designed to "exit" early and provide an answer if it reaches a high degree of confidence, avoiding unnecessary processing.


Study/Source

Year

Key Findings and Impact

Duke University, Nicholas Institute

2025

Found U.S. power grids could add nearly 100 GW of new flexible load if data centers curtailed their usage an average of 0.5% of the time annually, with an average curtailment time of about two hours. This demonstrates a significant, untapped potential for integrating large loads without costly grid upgrades.

Rocky Mountain Institute (RMI)

2025

States that if new data centers in the U.S. were to meet an annual load curtailment rate of 0.5%, it could make nearly 100 GW of new load available. The study emphasizes that temporal flexibility (demand response) offers benefits for both data centers (lower energy bills) and utilities (avoiding costly infrastructure upgrades).

Google Cloud Blog

2023

Describes a pilot program where Google used its "carbon-intelligent computing platform" for demand response. By shifting non-urgent tasks, they successfully reduced power consumption at their data centers during peak hours, supporting grid reliability in regions like Taiwan and Europe.

Emerald AI (Boston University)

2025

A technology field test in Phoenix, Arizona, showed that Emerald AI's software could reduce a data center's power usage by 25% during a period of peak electricity demand while maintaining service level agreements. The study highlights the potential of AI-driven strategies to dynamically adjust power usage and transform data centers into "virtual power plants."

174 Power Global

2025

Discusses how smart grid integration allows data centers to participate in demand response programs. It notes that facilities can shift computational workloads based on energy availability and cost, for example, by increasing processing for non-time-sensitive tasks during periods of high renewable energy generation.


As racks become denser with high-performance GPUs, liquid cooling systems can be up to 30 percent more energy-efficient than air cooling.


Separating the hot air exhausted by servers from the cold air intake also helps. By using "hot aisle/cold aisle" layouts with containment panels or curtains, data centers can prevent the air from mixing, allowing cooling systems to run less frequently.


“Free Cooling” in colder climates takes advantage of favorable outdoor temperatures. A data center can use outside air or water to cool the facility, bypassing mechanical chillers and significantly reducing energy consumption. 


Optimized Uninterruptible Power Supply (UPS) systems can reduce electrical losses by bypassing certain components when utility power is stable.


Data centers additionally can generate their own power using on-site solar, fuel cells, or battery storage. This keeps the data centers “off the grid” during peak demand on electrical utility networks. 


Server power capping limits the maximum power a server can draw, preventing  over-provisioning.


The point is that there always are multiple ways data centers can optimize their power usage to reduce electrical utility demand.


No comments:

Yes, Follow the Data. Even if it Does Not Fit Your Agenda

When people argue we need to “follow the science” that should be true in all cases, not only in cases where the data fits one’s political pr...