Thursday, September 25, 2025

Overinvestment is Typical for Any Big New Technology, But How Much AI Exposure is There?

Oracle  is lining up $38 billion for data center investments in two states tied to its “Project Stargate joint venture with OpenAI and SoftBank to build as much as $500 billion worth of new facilities supporting artificial intelligence operations over a four year period. 


The immediate $38 billion will support new facilities in Wisconsin and Texas that will be developed by Vantage Data Centers. Critics have argued that Project Stargate, when announced, was unfunded. 

The latest package is a step in that direction. 


Of course, there eventually will be concern about the payback from those investments, and investors already seem to be having an AI allergy at the moment. 


To a large extent, payback concerns center on the huge amount of depreciation the investments will trigger. One projection suggests that AI data centers to be built in 2025 will incur around $40 billion in annual depreciation, while generating only a fraction of that in revenue initially. 


The argument is that this massive depreciation, driven by the rapid obsolescence of specialized AI hardware, will outpace the revenue generation for years. 


Some might recall similar sorts of thinking in the first decade of the 21st century as cloud-based enterprise software began to replace traditional software licenses as a revenue model. Then, as now, there was concern about the relatively-low profit margin of cloud-delivered services, compared to use of traditional per-seat site licenses. 


That argument eventually was settled in favor of cloud delivery, however. So optimists will continue to argue that the financial return from massive AI data center investments will emerge, despite the high capital investment and equally-high depreciation impact on financial performance in the short term. 


But skeptics will likely continue to argue that, for processing-intensive AI operations, there is essentially no “long term,” as foundational chip investments must be continually refreshed with the latest versions, mirroring consumer experience with smartphone generations, for example, or mobile service provider experience with their infrastructure (2G, 3G, 4G, 5G and continuing). 


Among the issues is whether key graphics processing unit generations really need to be replaced every three or five years (possibly up to six years). Longer useful life means lower annual depreciation cost. 


But optimists expect demand will grow to more than match investments. 


Metric

Current Demand (2025 Est.)

Future Demand (2030 Est.)

Global AI Market Value

~$390 - $450 billion

~$1.3 - $1.8 trillion

Global Data Center Power Demand

~55 GW (14% from AI)

~122 GW (27% from AI)

Total AI-Related Capital Expenditure

~$200 billion/year

~$1 trillion/year

Required GPU Compute (Exaflops)

Low single-digit Exaflops

Hundreds of Exaflops

Demand-Supply Balance

Significant shortage

Supply-demand balance may be reached, but with continued investment pressure


Some cite the dot-com over-investment in optical fiber transport facilities as the basis for concern about AI data center investment. But there already seem to be key differences. The speculative investment in optical transport was based in large part on the expected survival and success of the many emerging “internet” firms. 


That did not happen when most of the startups failed. Also, there were some key instances of accounting fraud where firms were booking orders or revenue that did not actually exist. 


Depreciation schedules affect some capital-intensive businesses in a significant way.  


Venture capitalist David Cahn has estimated that for AI investments to be profitable, companies need to generate roughly four dollars in revenue for every dollar spent on capex. 


In the enterprise or business space, subscriptions seem the logical revenue model. In the consumer space, it is more likely that advertising will develop as the revenue model. But the issue is when that will happen. 


But there are other, more prosaic issues, such as the impact of depreciation on reported profitability. 

Historically, hyperscalers depreciated servers over three years.


In 2020 they started to extend server depreciation from three years  to four years. That might be deemed a simple recognition that better software enables existing hardware to provide value over longer time periods. 


As a practical matter, that front loads profits as training and inference revenues are booked before a significant amount of  depreciation expenses are recorded. 


In 2021, the hyperscalers investing in AI servers  further extended useful server life to five years and the useful life of networking gear from to six years, citing internal efficiency improvements that ‘lower stress on the hardware and extend the useful life’​.


Between 2021-2022, Microsoft, Alphabet and Meta followed suit, collectively lifting the useful lives of their server equipment to four years. In the year following, Microsoft and Alphabet further extended the depreciable lives for server equipment to six years, and Meta to five years.


It is too early to know which side of the debate will prove correct. 


Every Important New Automation Technology Causes Some Job Losses: Get Over It

It is not unusual for enterprise leaders to suggest artificial intelligence will lead to some job losses. In fact, it would be difficult to think of any instances of work automation that have not led to job losses in traditional settings. 


The Industrial Revolution provides some of the most dramatic examples of automation's effects. The invention of the steam engine and mechanized looms displaced countless textile artisans, leading to the Luddite movement where workers protested by destroying machinery. Later, the widespread adoption of the internal combustion engine and tractors decimated agricultural jobs and occupations tied to horses, like blacksmiths and stable hands.


In the 20th century, automated telephone switchboards replaced manual telephone operators, and the introduction of the Automated Teller Machine (ATM) reduced the number of bank tellers needed for routine transactions. More recently, self-service kiosks and online shopping have lessened the demand for cashiers and retail workers.


Technology

Replaced Human Role

How It Replaced the Role

Mechanized Loom

Textile Artisan (Weaver)

Automated the process of weaving, allowing a single machine to produce fabric much faster than a human.

Tractor/Harvester

Agricultural Laborer

Automated the manual and animal-powered tasks of farming, such as plowing and harvesting.

Telephone Switchboard

Telephone Operator

Automated the connection of phone calls, eliminating the need for human operators to manually plug lines.

Automated Teller Machine (ATM)

Bank Teller

Automated routine banking tasks like cash withdrawals and deposits, reducing the need for tellers.

Self-Checkout Kiosk

Cashier/Retail Worker

Automated the process of scanning and paying for goods in a retail environment.

Robotic Assembly Line

Factory Worker

Automated repetitive and dangerous tasks in manufacturing, such as welding and lifting heavy parts.

GPS & Digital Maps

Navigator/Pilot

Automated navigation and route planning, reducing the need for human expertise in these areas.



The new AI at Work Report 2025 published by Indeed suggests that more than a quarter (26 percent) of jobs posted on Indeed in the past year could be “highly” transformed by generative artificial intelligence apps. 


Some 54 percent of jobs are likely to be “moderately” transformed.


The study suggests 46 percent of skills in a typical U.S. job posting are poised for “hybrid transformation.” Human oversight will remain critical when applying these skills, but GenAI can already perform a significant portion of routine work.


As you might guess, software development and other cognitive functions are most likely to be affected, while jobs with high human contact, emotional intelligence or physical elements will be least affected. 


source: HiringLab.org 


Consider nursing, which is relatively immune from wholesale substitution effects. 


source: HiringLab.org 


In contrast, many more of the software development functions are likely to be affected. 


source: HiringLab.org 


Of the close to 3,000 requirements analyzed, the two dimensions that most directly determine task

transformation are:

• Problem-solving ability (cognitive reasoning, applied knowledge, and practical judgment)

• Physical necessity (physical execution, such as home construction, home repairs, plumbing and electrical work)


We might guess that the effects will extend as AI is embodied in more machines, with robotaxis and autonomous driving vehicles providing a good example. 


Tuesday, September 23, 2025

Embodied AI Will Extend the Range of Jobs and Functions AI can Displace

The new AI at Work Report 2025 published by Indeed suggests that more than a quarter (26 percent) of jobs posted on Indeed in the past year could be “highly” transformed by generative artificial intelligence apps. 


Some 54 percent of jobs are likely to be “moderately” transformed.


The study suggests 46 percent of skills in a typical U.S. job posting are poised for “hybrid transformation.” Human oversight will remain critical when applying these skills, but GenAI can already perform a significant portion of routine work.


As you might guess, software development and other cognitive functions are most likely to be affected, while jobs with high human contact, emotional intelligence or physical elements will be least affected. 


source: HiringLab.org 


Consider nursing, which is relatively immune from wholesale substitution effects. 


source: HiringLab.org 


In contrast, many more of the software development functions are likely to be affected. 


source: HiringLab.org 


Of the close to 3,000 requirements analyzed, the two dimensions that most directly determine task

transformation are:

• Problem-solving ability (cognitive reasoning, applied knowledge, and practical judgment)

• Physical necessity (physical execution, such as home construction, home repairs, plumbing and electrical work)


We might guess that the effects will extend as AI is embodied in more machines, with robotaxis and autonomous driving vehicles providing a good example.


MIT AI Report is Widely Misinterpreted

Much has been made of a study suggesting 95 percent of enterprises deploying artificial intelligence are not seeing a return on investment.


There’s just one glaring problem: the report points out that just five percent of those entities have AI in a “production” stage. The rest are pilots or limited early deployments. 


That significant gap between AI experimentation and successful, large-scale deployment arguably explains most of the sensationalized claim that “only five percent of enterprises” are seeing return on AI investment. 


It would be much more accurate to say that most enterprises have not yet deployed AI at scale, and therefore we cannot yet ascertain potential impact. 


Limited deployments or trials often do not integrate into core business workflows or provide the necessary scale to demonstrate significant financial impact, leading to the perception of widespread failure. And, in any case, up to 70 percent of all information technology projects fail to produce expected results. 


So it would be entirely normal for AI projects to fail much more often than they succeed. 


The report notes that “despite $30 to $40 billion in enterprise investment into GenAI, this report uncovers a surprising result in that 95 percent of organizations are getting zero return.”


But one has to dig just a bit deeper to make sense of those figures. The report notes that “just five percent” of integrated AI pilots are extracting millions in value, while the vast majority remain stuck with no measurable P&L impact.”


The report says “tools like ChatGPT and Copilot are widely adopted,” and that  “over 80 percent of organizations have explored or piloted them, and nearly 40 percent report deployment.”


But then come the important caveats. Those use cases “primarily enhance individual productivity, not P&L performance,” the report says.


Meanwhile, enterprise-grade systems, evaluated by 60 percent of organizations, have only had about 20 percent in a “pilot stage.” 


Most tellingly, “just five percent” or surveyed entities have enterprise AI systems in full production and use, the report says. 


The real issue is the percentage of firms that have fully deployed AI in a business process and are seeing return on investment. The report does not address that issue, but many of us might expect failure rates of up to 70 percent for those deployments. 


The point is that many, if not most, interpretations of the report’s data are off the mark. The study does not show enterprise AI use cases are not producing ROI 95 percent of the time. The report shows that most entities have not yet deployed at scale.


So, of course measurable returns are not available. One cannot measure the impact of an innovation one has not yet deployed at scale.


Monday, September 22, 2025

"All of the Above" Helps AI Data Centers Reduce Energy Demand

Just as electrical utilities use rate differentials to shift consumer workloads to off-peak hours, so data centers supporting artificial intelligence jobs can use workload shaping to shift jobs to off-peak hours. 


"Demand-side" management options are just as important as “supply side” increases in energy infrastructure. These demand management measures include:

  • optimizing hardware

  • improving software and workload management

  • enhancing physical infrastructure.


By identifying which AI tasks are time-sensitive (such as real-time inference for a search engine) versus which are not (training a new large language model), data centers can dynamically shift computational loads. 


Non-critical tasks can be paused or slowed down during peak grid demand hours and resumed when electricity is cheaper, cleaner, or more abundant. 


In a related way, workloads can be scheduled to run when the local grid's energy mix is dominated by renewable sources like solar or wind, which can reduce overall consumption.


The AI models themselves can be designed to be more efficient.


Techniques such as quantization and pruning reduce a model's size and the number of calculations required without significantly compromising accuracy. For example, by converting a model's parameters from 32-bit to 8-bit, its energy needs can be drastically reduced.


For certain tasks, an AI model can be designed to "exit" early and provide an answer if it reaches a high degree of confidence, avoiding unnecessary processing.


Study/Source

Year

Key Findings and Impact

Duke University, Nicholas Institute

2025

Found U.S. power grids could add nearly 100 GW of new flexible load if data centers curtailed their usage an average of 0.5% of the time annually, with an average curtailment time of about two hours. This demonstrates a significant, untapped potential for integrating large loads without costly grid upgrades.

Rocky Mountain Institute (RMI)

2025

States that if new data centers in the U.S. were to meet an annual load curtailment rate of 0.5%, it could make nearly 100 GW of new load available. The study emphasizes that temporal flexibility (demand response) offers benefits for both data centers (lower energy bills) and utilities (avoiding costly infrastructure upgrades).

Google Cloud Blog

2023

Describes a pilot program where Google used its "carbon-intelligent computing platform" for demand response. By shifting non-urgent tasks, they successfully reduced power consumption at their data centers during peak hours, supporting grid reliability in regions like Taiwan and Europe.

Emerald AI (Boston University)

2025

A technology field test in Phoenix, Arizona, showed that Emerald AI's software could reduce a data center's power usage by 25% during a period of peak electricity demand while maintaining service level agreements. The study highlights the potential of AI-driven strategies to dynamically adjust power usage and transform data centers into "virtual power plants."

174 Power Global

2025

Discusses how smart grid integration allows data centers to participate in demand response programs. It notes that facilities can shift computational workloads based on energy availability and cost, for example, by increasing processing for non-time-sensitive tasks during periods of high renewable energy generation.


As racks become denser with high-performance GPUs, liquid cooling systems can be up to 30 percent more energy-efficient than air cooling.


Separating the hot air exhausted by servers from the cold air intake also helps. By using "hot aisle/cold aisle" layouts with containment panels or curtains, data centers can prevent the air from mixing, allowing cooling systems to run less frequently.


“Free Cooling” in colder climates takes advantage of favorable outdoor temperatures. A data center can use outside air or water to cool the facility, bypassing mechanical chillers and significantly reducing energy consumption. 


Optimized Uninterruptible Power Supply (UPS) systems can reduce electrical losses by bypassing certain components when utility power is stable.


Data centers additionally can generate their own power using on-site solar, fuel cells, or battery storage. This keeps the data centers “off the grid” during peak demand on electrical utility networks. 


Server power capping limits the maximum power a server can draw, preventing  over-provisioning.


The point is that there always are multiple ways data centers can optimize their power usage to reduce electrical utility demand.


Has AI Use Reached an Inflection Point, or Not?

As always, we might well disagree about the latest statistics on AI usage. The proportion of U.S. employees who report using artificial inte...