Neubauer Coporation Getting your Trinity Audio player ready... |
A Generative AI system might use around 33 times more energy than machines running task-specific software, according to a recent study by Dr Luccioni and colleagues. The work has been peer-reviewed but is yet to be published in a journal.
“The cloud,” says Dr Luccioni. “You don’t think about these huge boxes of metal that heat up and use so much energy.”
Much of the nervousness over AI’s energy use is due to the recent rise of large language models. The process of training these models, which entails ingesting huge amounts of data, has a high impact on its own. Back in 2019, researchers determined that training a large AI model produced about five times the lifetime emissions of the average car. That sounds catastrophic, but perhaps less so when you consider how many more cars there are on the road than AI models being trained every day.
The world’s data centres are using ever more electricity. In 2022, they gobbled up 460 terawatt hours of electricity, and the International Energy Agency (IEA) expects this to double in just four years. Data centres could be using a total of 1,000 terawatts hours annually by 2026. “This demand is roughly equivalent to the electricity consumption of Japan,” says the IEA. Japan has a population of 125 million people.
Luccioni worked on a study that looked at the energy use of different tasks performed by AI models. It found that, in order to generate 1,000 images, “the least efficient image generation model uses as much energy as 522 smartphone charges (11.49 kWh), or around half a charge per image generation.” (One caveat was that there was a lot of variation across the image-generation models, depending on the size of the image generated.) Text-based tasks were much more efficient, though they still add up.
While that’s a significant amount of energy, each individual use isn’t so much a problem on its own, and how concerning you deem it may depend on your frame of reference. A separate group of researchers, predominantly from the University of California, compared the carbon emissions of using AI for writing and illustrating tasks versus having humans perform them. The AI systems actually yielded fewer emissions than humans when factoring in how long it takes a human to write a page or create an illustration and the emissions produced by running a computer over that period.
It’s arguably not a perfect comparison. An AI system could quickly write a page of copy or produce an image, but the results may then require a human to go back and edit. Still, if it saves enough time, the point stands that it could compare favourably with a human working at an electricity-powered computer, probably saving their work to the cloud, which is really just another name for remote servers. (The fact that AI could displace workers and that popular generative models are trained on creative work without consent are separate, if no less serious, matters.)
At data centres, huge volumes of information are stored for retrieval anywhere in the world – everything from your emails to Hollywood movies. The computers in those faceless buildings also power AI and cryptocurrency. They underpin life as we know it.
But some countries know all too well how energy hungry these facilities are. There is currently a moratorium preventing the construction of new data centres in Dublin. Nearly a fifth of Ireland’s electricity is used up by data centres, and this figure is expected to grow significantly in the next few years – meanwhile Irish households are reducing their consumption.
The boss of National Grid said in a speech in March that data centre electricity demand in the UK will rise six-fold in just 10 years, fuelled largely by the rise of AI. National Grid expects that the energy required for electrifying transport and heat will be much larger in total, however.
Utilities firms in the US are beginning to feel the pressure, says Chris Seiple at Wood Mackenzie, a research firm.
“They’re getting hit with data centre demands at the exact same time as we have a renaissance taking place – thanks to government policy – in domestic manufacturing,” he explains. Lawmakers in some states are now rethinking tax breaks offered to data centre developers because of the sheer strain these facilities are putting on local energy infrastructure, according to reports in the US.
Mr Seiple says there is a “land grab” going on for data centre locations near to power stations or renewable energy hubs: “Iowa is a hotbed of data centre development, there’s a lot of wind generation there.”
Some data centres can afford to go to more remote locations these days because latency – the delay, usually measured in milliseconds, between sending information out from a data centre and the user receiving it – is not a major concern for increasingly popular Generative AI systems. In the past, data centres handling emergency communications or financial trading algorithms, for example, have been sited within or very near to large population centres, for the absolute best response times.
There is little doubt that the energy demands of data centres will rise in the coming years, but there is huge uncertainty over how much, stresses Mr Seiple.
Part of that uncertainty is down to the fact that the hardware behind generative AI is evolving all the time.
Tony Grayson is general manager at Compass Quantum, a data-centre business, and he points to Nvidia’s recently launched Grace Blackwell supercomputer chips (named after a computer scientist and a mathematician), which are designed specifically to power high-end processes including generative AI, quantum computing and computer-aided drug design.
Nvidia says that, in the future, a company could train AIs several times larger than those currently available in 90 days. It would require 8,000 of the previous generation of Nvidia chips and need a 15 megawatt electricity supply.
But the same work could be carried out in the same time by just 2,000 Grace Blackwell chips, and they would need a four megawatt supply, according to Nvidia.
That still ends up as 8.6 gigawatt hours of electricity consumed – roughly the same amount that the entire city of Belfast uses in a week.
“The performance is going up so much that your overall energy savings are big,” says Mr Grayson. But he agrees that power demands are shaping where data centre operators site their facilities: “People are going to where cheap power’s at.”
Dr Luccioni notes that the energy and resources required to manufacture the latest computer chips are significant.
Still, it is true that data centres have got more energy efficient over time, argues Dale Sartor, a consultant and affiliate of Lawrence Berkeley National Laboratory in the US. Their efficiency is often measured in terms of power usage effectiveness, or PUE. The lower the number, the better. State-of-the-art data centres have a PUE of around 1.1, he notes.
These facilities do still create significant amounts of waste heat and Europe is ahead of the US in finding ways of using that waste heat – such as warming up swimming pools – says Mr Sartor.
Bruce Owen, UK managing director at Equinix, a data centre firm, says, “I still think that the demand is going to grow further than that efficiency gain that we see.” He predicts that more data centres will be built with on-site power-generating facilities included. Equinix was denied planning permission for a gas-powered data centre in Dublin last year.
Mr Sartor adds that costs may ultimately determine whether Generative AI is worth it for certain applications: “If the old way is cheaper and easier then there’s not going to be much of a market for the new way.”
Meanwhile, data centres will be busy powering the rest of the digital world, too, including social media, e-commerce, streaming video, online gaming and more — all of them requiring energy. Ars Technica also noted that the amount of electricity consumed by PC gaming according to one 2018 study was nearly as great as the amount de Vries estimated AI would use in the next few years.
Dr Luccioni stresses, though, that people will need to clearly understand how the options in front of them differ in terms of energy efficiency. She is working on a project to develop energy ratings for AI.
“Instead of picking this GPT-derivative model that is very clunky and uses a lot of energy, you can pick this A+ energy star model that will be a lot more lightweight and efficient,” she says.
So should brands be concerned about AI’s impact? They absolutely should. But they should also be aware of the impact they’re having in other ways, too, because the environmental problems around AI aren’t totally unique to the technology. Just in running their daily operations they’re using energy and generating emissions. What matters more than just whether or not a company chooses to use AI is whether it’s even considering and trying to reduce its impact overall.