Yves here. This article provides a clear and accessible overview of the challenges posed by the rapid expansion of data centers, including a surprising new obstacle: a shortage of transformers.
By Michael Kern, newswriter and editor at Safehaven.com and Oilprice.com. Originally published at OilPrice
- Shifts in AI from CPUs to ultra-dense GPUs are leading to a surge in electricity demand that exceeds efficiency improvements, straining the electrical grid.
- The requirement for reliable, 24/7 power is slowing down coal plant retirements, increasing the use of natural gas, and driving the growth of data centers to regions with lower land costs, lax regulations, and underserved communities.
- Global shortages of transformers, along with limitations in essential minerals and infrastructure bottlenecks, suggest that computing power, rather than oil, might become the next resource nations seek to hoard.
Referring to “the Cloud” may be one of the best branding maneuvers ever. It conjures images of something light, airy, and almost insubstantial. Yet, if you visit places like Loudoun County, Virginia, or the desolate expanses of Altoona, Iowa, it’s clear that the Cloud is, in reality, a massive, noisy, and overheated factory.
We’ve been comforted by a narrative about the energy transition, celebrating the closure of coal plants and the construction of wind farms. We believed we were successfully uncoupling economic growth from carbon emissions.
For years, the tech industry managed to maintain a relative decoupling…
Moore’s Law allowed improvements in server efficiency to stay ahead of demand, enabling the internet to expand while energy usage rose slowly.
However, the rapid growth of AI has disrupted this balance. AI’s computational requirements are so intense that energy demand is now skyrocketing faster than efficiency can keep up.
We are witnessing a realignment with the laws of physics; the central narrative for the upcoming decade revolves less around supply and more around an unanticipated shift in demand: the thermodynamics of artificial intelligence.
The International Energy Agency (IEA) projects that by 2030, global electricity demand from data centers will double, equivalent to the total electricity consumption of a nation like Japan.
This raises pressing questions about the grid’s capacity; we are no longer debating if it can cope, but rather what is behind the unprecedented demand surge. The issues do not lie in software improvements or smarter algorithms; they are fundamentally about the intense physical requirements of data centers, which now need as much power as a small city.
The Thermodynamics of “Thinking”
To grasp why the grid is under pressure, we must examine the silicon technology.
Historically, our internet infrastructure relied on CPUs (Central Processing Units), serving as the efficient, reliable core of computing. However, Generative AI demands significantly more, opting for GPUs (Graphics Processing Units) like Nvidia’s H100.
Consider the implications for power consumption:
- Traditional server rack: Consumes around 5 to 10 kilowatts (kW).
- Modern AI rack (H100s/Blackwell): Draws between 50 to 100 kW.
This shift means we are no longer powering a single appliance but an entire community, all within the confines of a single metal box. Traditional air cooling methods are ineffective as air cannot efficiently dissipate the immense heat generated.
Consequently, data centers are adopting advanced cooling methods akin to what you would see in refineries, with liquid cooling systems directed straight toward the silicon to manage the extreme thermal output.
This approach, termed Direct-to-Chip (DTC) cooling, is becoming standard in leading-edge AI facilities, as it is crucial for sustaining the severe heat generated by chips like the H100.
Liquid cooling is more energy-efficient than traditional air conditioning methods. Although the chip itself still draws 100 kW, the overall cooling setup—comprising pumps and chillers—uses considerably less power than maintaining extensive ventilation systems. This solution is born from necessity.
The next evolution is Immersion Cooling, where entire server racks are submerged in non-conductive fluids, already being tested in pilot projects and specialized facilities.
This transition from airflow to specialized cooling systems reflects the industrialization of cognitive processes.
Similar to the industrial revolutions in textiles or steel production, this shift necessitates huge quantities of raw power and specialized materials. Such industrial demands challenge traditional renewable energy sources, like intermittent solar and wind, to provide the reliability needed.
When an AI training session costs millions, a mere 1% flicker in power supply poses an existential risk.
The Dirty Secret of the “Green” AI Boom
Today, prominent tech CEOs promote their “Net Zero” goals for 2030 during podcasts. Sure, many are investing in carbon credits.
Yet, the laws of physics remain unaltered by carbon offsets. The truth is that AI requires steady, round-the-clock energy supply, with a reliability rate of “five nines” (99.999%).
What provides that reliability? According to IEA data, coal still contributes about 30% of global data center energy needs, while in the U.S., over 40% is fulfilled by natural gas.
The irony is striking. After spending billions to phase out coal, the most advanced technology—AI—now provides it with renewed relevance.
In regions like Virginia and Kansas, utility companies are postponing the closure of coal plants; they cannot risk grid instability when a large-scale data center is activated.
The “future” relies on the “past.”
This urgent need for dependable baseload power, coupled with the immense energy appetite of contemporary facilities, is reshaping the American energy landscape. Investment flows toward opportunities with the least resistance, which currently leads through regions that lack the benefits of tech prosperity.
The New Geography of Power (and Inequality)
The insatiable energy demand is altering geographical patterns. We are witnessing a “K-shaped” infrastructure landscape emerge.
In the U.S., “Data Center Alley” in Northern Virginia is supposedly responsible for 70% of global internet traffic. However, the grid there is at capacity, with no new connections available for years.
As a result, investment is fleeing to areas with relaxed regulations and more affordable land, including Texas, Ohio, and Arizona.
This scenario presents complications. These facilities often become problematic neighbors. They generate noise, consume vast amounts of water for cooling, and escalate local utility costs.
A key aspect of Environmental Justice emerges here, as industrial sites are seldom established in affluent neighborhoods.
The NAACP’s “Fumes Across the Fence-Line” report reveals:
- African Americans are 75% more likely than white Americans to reside in “fence-line” communities adjacent to industrial facilities.
- A disproportionate number of fossil-fuel peaker plants, operational at times when data centers max out the grid, are sited in low-income neighborhoods and communities of color.
This reality correlates with higher asthma and respiratory issue rates.
While the “invisible wealth” stemming from AI stock increases benefits wealthy portfolios in cities like San Francisco and New York, the “visible decline”—pollution, water demands, and incessant cooling fan noise—is confined to communities that largely miss out on the benefits.
Even if a community is prepared to absorb these costs, the traditional infrastructure that once reliably supplied the electrical grid is now overwhelmed.
The question has shifted from where to locate data centers, to how to connect these massive, energy-consuming factories to the current grid. This process is severely hindered by a global shortage of critical non-digital hardware.
The Great Transformer Shortage
Let’s say you have the necessary funding, land, and permits. You still face a significant hurdle: the gear.
The delivery time for high-voltage transformers once averaged 12 months. Now, it stretches to 3 to 5 years.
We are attempting to modernize the electrical grid while everyone else is also looking to electrify cars and heat pumps, creating a fractured supply chain.
Moreover, we are depleting essential raw materials: copper, lithium, and neodymium for fan magnets.
Our reliance on China for the processing of these critical minerals intensifies the situation. As discussed in my “Data Center Guide,” we are beginning to understand that the digital economy is actually a material-driven one.
If China restricts exports of graphite or gallium, the Cloud’s expansion halts.
The “Trust Me, Bro” Efficiency Pitch
Silicon Valley’s counter-argument is rooted in the “Handprint” theory, which claims: Yes, training AI consumes significant energy, but its efficiency gains will compensate for it.
According to IEA models, AI has the potential to optimize logistics, manage smart grids, and cut down building energy use by 10-20%.
This is indeed a persuasive argument. If AI can enhance the efficiency of a truck fleet by 5%, that can offset more carbon emissions than the data center generates.
However, this is a long-term gamble against the immediate, unavoidable power demands.
Two primary issues are contributing to efficiency struggles:
- Training vs. Inference: Training an expansive model necessitates an enormous power surge over extended periods. While the trained model’s inference (its ability to answer queries) is cheaper per interaction, the skyrocketing overall volume transforms small energy inputs into a massive, ongoing drain.
- The Hardware Treadmill: High-end CPUs may last 5-7 years in a data center, but new AI GPUs become obsolete in just two years. This relentless cycle of replacing power-hungry H100s with even more demanding Blackwells prevents the carbon and raw materials tied to the silicon from being recouped over reasonable operational lifespans.
Currently, we are expending carbon with the hope of achieving efficiencies later. Although firms are working on “smarter” silicon and efficient ASICs for inference, this transition will not arrive in time to mitigate the ongoing surge in demand on the grid.
What Comes Next?
We are transitioning from an era defined by Generation Constraints to one dominated by Connection Constraints.
The most coveted asset now isn’t the H100 chip but a signed interconnection agreement with a utility provider. The “queue” to access the grid has become the new standard for privilege.
This shift will initiate several critical changes:
- Off-Grid AI: Technology giants may cease waiting for utility support. They might begin constructing their own Small Modular Nuclear Reactors (SMRs) or extensive solar farms with battery backing, effectively opting to operate independently.
- Sovereign Compute: Nations will recognize that “compute” is as vital a resource as oil. We may witness countries hoarding energy to sustain their AI models instead of exporting it.
- The Efficiency Wall: We’ll reach a tipping point where the cost of power renders brute-force AI training nonviable, prompting a shift to smarter chips (ASICs) and possibly paving the way for neuromorphic or photonic computing technologies.
The invisible hand of the market is revealing its cards, yet the laws of thermodynamics are challenging the bluff. The virtual arena relies on tangible energy, and for the first time in years, we realize that the concept of “unlimited data” was merely an ephemeral illusion.