Arm CEO warns AI’s power appetite could devour 25% of US electricity by 2030
Arm CEO Rene Haas cautions that if AI continues to get more powerful without boosts in power efficiency, datacenters could consume extreme amounts of electricity.
Haas estimates that while US power consumption by AI datacenters sits at a modest four percent, he expects the industry to trend towards 20 to 25 percent usage of the US power grid by 2030, per a report from the Wall Street Journal. He specifically lays blame at popular large language models (LLMs) such as ChatGPT, which Haas described as “insatiable in terms of their thirst.”
The Arm CEO isn’t alone in making this prediction. The International Energy Agency’s (IEA) Electricity 2024 report [PDF] expects power consumption for AI datacenters around the world to be ten times the amount it was in 2022. Part of the problem is that LLMs like ChatGPT require far more power than traditional search engines like Google. The IEA estimates that one ChatGPT request consumes almost ten times as much power as a Google search.
https://www.theregister.com/2024/04/09/ai_datacenters_unsustainable/