It’s hard to point to any impressive human advance that didn’t have an equally impressive cost to nature. From the smoke-filled streets of the first industrial cities to the Great Pacific Garbage Patch, history has many examples of epic ecological catastrophes. While a few years ago everyone’s eyes were on energy-hungry blockchain mining rigs, today […]
It’s hard to point to any impressive human advance that didn’t have an equally impressive cost to nature. From the smoke-filled streets of the first industrial cities to the Great Pacific Garbage Patch, history has many examples of epic ecological catastrophes.
While a few years ago everyone’s eyes were on energy-hungry blockchain mining rigs, today no one can ignore the rapidly growing impact AI has on the planet.
Deloitte predicts that the amount of electricity consumed by data centres worldwide will reach over 1,000 terawatt-hours (TWh) by 2030 and 2,000 TWh by 2050, making up nearly 3% of the world’s total electricity consumption. The culprit? As you can probably guess, Large Language Models (LLMs) require tons of computing power both for training and operations. And while the tools of reducing this impact are there, their implementation is less than straightforward.
The root cause(s) of AI-led environmental impact
While the general public is probably troubled by headlines like “one ChatGPT request consumes 10 times more energy than one Google search” or “generating one AI image costs as much in energy as half a smartphone charge”, the consumer-facing side of AI models is just the tip of the iceberg. And if the reduction in AI-related energy consumption is seen as a pressing issue, we should start at the source. Namely, at the training stage and data consumption.
The massive energy cost is mostly due to the fact that the majority of AI models (including those by OpenAI and DeepMind) on the market today are stochastic in nature. In other words, the results they generate are based on complicated algorithmic calculations that need a lot of power to run.
Couple that with the high appetite for data these models have during training, and you can see how that compounds and creates a toll on the environment. To put things in perspective, training GPT-3 produced 552 metric tons of CO₂, equivalent to driving a car for over 2 million kilometers.
The winding road to sustainable AI
Environmental shifts that go beyond buying carbon credits or planting a few trees are hard for any industry. In the case of AI developers, the external pressure to prioritise green solutions comes from investors and governments.
For instance, a 2023 survey by McKinsey shows that 35% of investors would consider decreasing their exposure to tech companies due to Environmental, Social, and Governance (ESG) concerns, with the ‘E’ part playing an important role. Does this mean that performance must be sacrificed in return? Not necessarily, as some viable solutions offer a solid compromise between the two.
First, GPU efficiency must be considered a hygiene factor in the AI industry. This does not mean that companies need to get rid of their existing rigs. A research paper presented at the USENIX Symposium on Networked Systems Design and Implementation explored how a more deliberate use of the GPU power limit caps can reduce energy consumption by up to 75%.
Alongside GPU optimisation, we have another low-hanging fruit that too few AI companies consider. Namely, a more efficient approach to data collection and management, as parsing terabytes of redundant, low-quality data takes its toll on energy consumption, too.
Will tomorrow’s data centres be nuclear powered?
Unfortunately, ‘redundant, low-quality data’ is a synonym for a lot of poorly sourced data from the web. When done right, web scraping can provide clean and efficient data sets, but not all providers of such services prioritise quality over sheer quantity. As the digital economy expands its data footprint by 40% every year, this problem is going to become more acute.
Given that data being generated by today’s increasingly digitised economy is predicted to reach a truly unfathomable 163 trillion gigabytes in 2025, optimising data sets coming from the web will be more important than ever. On the positive side, new algorithmic approaches, such as probabilistic programming, allow for the use of smaller datasets instead of relying on massive-scale data processing.
Just as data quality varies, so does the environmental impact of different energy sources employed in AI business. While we are yet to see the consequences of the US withdrawing from the Paris Agreement in 2025, there are plenty of countries that prioritise green energy on the other side of the pond.
Relocating data centres to places with abundant (and potentially cheaper) green energy can make a massive difference. Take Google, whose Finland-based data centre runs on 97% carbon-free energy, which would have been an impossible feat to achieve in countries where fossil fuels are still king.
Finally, federated learning emerges as a viable alternative to centralised AI training. Unlike more traditional machine learning, federated learning distributes the computational burden across multiple sources — from servers to smartphones, desktop computers, and so on.
A study by the University of Cambridge showed that a federated learning setup, if configured properly, can emit significantly less CO₂ than centralised training methods.
Good intentions and declarations are not enough
Having outlined several technical solutions for a greener AI development, it is important to note that the path to their implementation is anything but straightforward. In an industry where the fear of a “bursting AI bubble” is on everyone’s mind, going green might not be seen as a priority.
First, there’s the harsh reality of diminishing returns. As AI systems grow more complex, their ROI is, unfortunately, not keeping pace with expectations, and most AI companies are still in the red. For instance, in 2024, OpenAI had accumulated approximately $8 billion in expenses, which put it on the brink of bankruptcy in July.
Recent projections suggest that these losses could go up to $14 billion by 2026. For reference, the company’s revenue for 2024 stood at just $3.7 billion. Looking at other major players, the situation is not much brighter.
Another strong barrier is the fear of falling behind, which is felt acutely in the West due to China’s increasingly aggressive push into AI development. Thus, prioritising environmental concerns might be seen as a business disadvantage.
Going from performance-first “Red AI” to Green AI might also be technically challenging. Balancing performance and energy efficiency takes more than just moving servers to Finland. Ensuring high performance is still key, unless you want a model that hallucinates or works slower than its competitors.
Big Tech enables climate disinformation, study finds
Besides, the technical talent needed to implement probabilistic algorithms and optimise data management is hard to find and costly.
Finally, the industry has yet to adopt a standardised way of defining and measuring AI sustainability. While initiatives like MLPerf Green have made some progress in setting common benchmarks, without consensus among the major players, there is still a lot of room for greenwashing.
The benefits of starting at the source
Despite all challenges, the industry must move forward. The United Nations Environment Programme (UNEP) has recently outlined a comprehensive framework that can hopefully strike a balance between what businesses and the planet need.
According to UNEP, countries should introduce standardised procedures for measuring the environmental impact of AI, while governments should make it a requirement for companies to disclose the direct environmental consequences of AI-based products and services.
For their part, tech companies have to fine-tune their AI algorithms for efficiency, while reducing their overall demand for energy. On the national level, governments should incentivise companies to use green energy for data centres (perhaps via tax cuts). At the same time, AI-related policies should become a standard part of broader environmental regulations.
If any of these sound like lofty goals, they are. However, there is little preventing AI companies from taking the first step forward. Among all the solutions we discussed, optimising data quality and data flow is probably going to bring the biggest and quickest business pay-offs.
And in the long run, investing in better quality data will also result in more competitive AI models. Hopefully, in 2025 and beyond, the reduction of the environmental AI footprint will start at the source of the problem by finally giving more attention to structured, targeted data for the full cycle of AI operations.