The Hindu (Mumbai)

AI has a big and growing carbon footprint, but algorithms can help

Training GPT-3, the precursor AI system to the current ChatGPT, generated 502 metric tonnes of carbon, which is equivalent to driving 112 petrol cars for a year. GPT-3 further emits 8.4 tonnes of carbon dioxide annually due to inference. The requiremen­ts

- Shirin Dora

AI could also help us in tackling the climate crisis. However, when we consider energy needs it becomes clear that the technology is as much a part of the problem as a solution

Given the huge problemsol­ving potential of artificial intelligen­ce (AI), it wouldn’t be farfetched to think that AI could also help us in tackling the climate crisis. However, when we consider the energy needs of AI models, it becomes clear that the technology is as much a part of the climate problem as a solution.

The emissions come from the infrastruc­ture associated with AI, such as building and running the data centres that handle the large amounts of informatio­n required to sustain these systems.

But different technologi­cal approaches to how we build AI systems could help reduce its carbon footprint. Two technologi­es in particular hold promise for doing this: spiking neural networks and lifelong learning.

The lifetime of an AI system can be split into two phases: training and inference. During training, a relevant dataset is used to build and tune – improve – the system. In inference, the trained system generates prediction­s on previously unseen data.

For example, training an AI that’s to be used in selfdrivin­g cars would require a dataset of many different driving scenarios and decisions taken by human drivers.

After the training phase, the AI system will predict effective manoeuvres for a selfdrivin­g car. Artificial neural networks (ANN), are an underlying technology used in most current AI systems.

They have many different elements to them, called parameters, whose values are adjusted during the training phase of the AI system. These parameters can run to more than 100 billion in total.

While large numbers of parameters improve the capabiliti­es of ANNs, they also make training and inference resourcein­tensive processes. To put things in perspectiv­e, training GPT3 (the precursor AI system to the current ChatGPT) generated 502 metric tonnes of carbon, which is equivalent to driving 112 petrol powered cars for a year.

GPT3 further emits 8.4 tonnes of CO2 annually due to inference. Since the AI boom started in the early 2010s, the energy requiremen­ts of AI systems known as large language models — the type of technology that’s behind ChatGPT — have gone up by a factor of 300,000.

With the increasing ubiquity and complexity of AI models, this trend is going to continue, potentiall­y making AI a significan­t contributo­r of CO2 emissions. In fact, our current estimates could be lower than AI’s actual carbon footprint due to a lack of standard and accurate techniques for measuring AIrelated emissions.

Spiking neural networks

The previously mentioned new technologi­es, spiking neural networks (SNNs) and lifelong learning (L2), have the potential to lower AI’s everincrea­sing carbon footprint, with SNNs acting as an energyeffi­cient alternativ­e to ANNs.

ANNs work by processing and learning patterns from data, enabling them to make prediction­s. They work with decimal numbers. To make accurate calculatio­ns, especially when multiplyin­g numbers with decimal points together, the computer needs to be very precise. It is because of these decimal numbers that ANNs require lots of computing power, memory and time.

This means ANNs become more energyinte­nsive as the networks get larger and more complex. Both ANNs and SNNs are inspired by the brain, which contains billions of neurons (nerve cells) connected to each other via synapses.

Like the brain, ANNs and SNNs also have components which researcher­s call neurons, although these are artificial, not biological ones. The key difference between the two types of neural networks is in the way individual neurons transmit informatio­n to each other.

Neurons in the human brain communicat­e with each other by transmitti­ng intermitte­nt electrical signals called spikes. The spikes themselves do not contain informatio­n. Instead, the informatio­n lies in the timing of these spikes. This binary, allornone characteri­stic of spikes (usually represente­d as 0 or 1) implies that neurons are active when they spike and inactive otherwise.

This is one of the reasons for energy efficient processing in the brain.

Just as Morse code uses specific sequences of dots and dashes to convey messages, SNNs use patterns or timings of spikes to process and transmit informatio­n. So, while the artificial neurons in ANNs are always active, SNNs consume energy only when a spike occurs.

Otherwise, they have closer to zero energy requiremen­ts. SNNs can be up to 280times more energy efficient than ANNs.

My colleagues and I are developing learning algorithms for SNNs that may bring them even closer to the energy efficiency exhibited by the brain. The lower computatio­nal requiremen­ts also imply that SNNs might be able to make decisions more quickly.

These properties render SNNs useful for a broad range of applicatio­ns, including space exploratio­n, defence and selfdrivin­g cars because of the limited energy sources available in these scenarios.

L2 is another strategy for reducing the overall energy requiremen­ts of ANNs over the course of their lifetime that we are also working on.

Training ANNs sequential­ly (where the systems learn from sequences of data) on new problems causes them to forget their previous knowledge while learning new tasks.

ANNs require retraining from scratch when their operating environmen­t changes, further increasing AIrelated emissions.

L2 is a collection of algorithms that enable AI models to be trained sequential­ly on multiple tasks with little or no forgetting.

L2 enables models to learn throughout their lifetime by building on their existing knowledge without having to retrain them from scratch.

The field of AI is growing fast and other potential advancemen­ts are emerging that can mitigate the energy demands of this technology. For instance, building smaller AI models that exhibit the same predictive capabiliti­es as that of a larger model.

Advances in quantum computing — a different approach to building computers that harnesses phenomena from the world of quantum physics — would also enable faster training and inference using ANNs and SNNs.

The superior computing capabiliti­es offered by quantum computing could allow us to find energyeffi­cient solutions for AI at a much larger scale.

The climate change challenge requires that we try to find solutions for rapidly advancing areas such as AI before their carbon footprint becomes too large.

(Shirin Dora is lecturer, computer science, Loughborou­gh University. This article is republishe­d from The Conversati­on.)

 ?? NAT/UNSPLASH ?? AI systems make huge demands on energy resources.
NAT/UNSPLASH AI systems make huge demands on energy resources.
 ?? ??

Newspapers in English

Newspapers from India