Home / Chroniques / Is AI doomed to be an energy drain?
Modern data center with servers
π Digital π Energy

Is AI doomed to be an energy drain ?

Enzo Tartaglione
Enzo Tartaglione
Associate Professor at Télécom Paris (IP Paris)
Key takeaways
  • Whether it’s training a model or making extrapolations, generative AI is energy intensive.
  • The energy consumption is due to its extrapolations constantly increasing – Amazon Web Services estimates that 90% of cloud machine learning demand comes from this.
  • At Télécom Paris (IP Paris), a specialised research chair is looking into how to reconcile the rise of AI with energy constraints without sacrificing its potential.
  • One of the proposed solutions is to optimise models by dividing them into a multitude of “experts” capable of activating themselves according to the task at hand.
  • Improving the energy efficiency of GAI would not only benefit the environment but also have a positive economic impact for those developing this type of tool.

Gene­ra­tive AI models, such as OpenAI’s GPT‑4, are pre­sen­ted as all-pur­pose tools. They include a signi­fi­cant num­ber of para­me­ters – now num­be­ring in the bil­lions – that enable them to per­form any type of task. This plu­ra­li­ty of uses, which leads to com­plexi­ty issues, makes these models “in need of opti­mi­sa­tion”, accor­ding to Enzo Tar­ta­glione, resear­cher and senior lec­tu­rer at Télé­com Paris (IP Paris). This com­plexi­ty also implies consi­de­rable ener­gy consump­tion.

“Even for an extre­me­ly simple que­ry, AI will tend to use all the resources at its dis­po­sal to respond, without exclu­ding those that are not use­ful. This leads to ener­gy waste, and it is real­ly some­thing we need to opti­mise.” This ener­gy consump­tion, esti­ma­ted at around 2% of glo­bal consump­tion in 2024, is dri­ving research towards an alter­na­tive approach : ener­gy efficiency.

From training to use

Ope­nAI has made an extre­me­ly resource-inten­sive lan­guage model avai­lable via ser­vers. This obser­va­tion led resear­chers to dis­tin­guish bet­ween the resource consump­tion of model trai­ning and that of its infe­rences, i.e. its use. Although the ener­gy consump­tion of trai­ning is signi­fi­cant – approxi­ma­te­ly 1,287 MWh for GPT‑3, and bet­ween 10,000 and 30,000 MWh esti­ma­ted for GPT‑4 – its impact is one-off. The impact of infe­rence, on the other hand, depends on the num­ber of users, which is gro­wing constant­ly. A 2021 stu­dy1 esti­mates that “bet­ween 80 and 90% of machine lear­ning work­load at NVIDIA comes from infe­rence. Ama­zon Web Ser­vices esti­mates that 90% of cloud demand for machine lear­ning is inference.”

Some resear­chers believe that a balance needs to be found bet­ween a model’s ener­gy consump­tion and the task it is requi­red to per­form. If a model is used to dis­co­ver a drug or advance research – which it is capable of doing – the car­bon foot­print will be easier to accept. Howe­ver, today, these models can be used for all kinds of tasks, making mil­lions of infe­rences through the various requests made of them at the same time. 

At Télé­com Paris (IP Paris), the Data Science and Arti­fi­cial Intel­li­gence for Digi­ta­li­sed Indus­try and Ser­vices chair focuses on seve­ral chal­lenges : how to recon­cile the rise of AI and its ener­gy constraints without sacri­fi­cing its poten­tial. “We are explo­ring issues of fru­ga­li­ty (editor’s note : see­king to “do more with less” and with grea­ter res­pect for the envi­ron­ment), but also sus­tai­na­bi­li­ty (editor’s note : mee­ting the needs of present gene­ra­tions without com­pro­mi­sing those of future gene­ra­tions),” adds Enzo Tar­ta­glione. “There is a real ques­tion in the choice of appli­ca­tions, because we can­not sim­ply point the fin­ger at AI and say it is a bad thing. Toge­ther with col­leagues, we are star­ting work on gene­ra­ting mate­rials for sto­ring hydro­gen. This is also some­thing that AI can offer as a solution.”

Espe­cial­ly since the models we can all use on our mobile phones require com­mu­ni­ca­tion with a ser­ver. “We need to be aware of the cost of trans­por­ting infor­ma­tion, espe­cial­ly when it’s bidi­rec­tio­nal,” insists the resear­cher. “There is the­re­fore a great neces­si­ty to desi­gn models that can be used local­ly, limi­ting the need for com­mu­ni­ca­tion with an exter­nal ser­ver. Howe­ver, we are tal­king about models with bil­lions of dif­ferent para­me­ters. This requires too much memo­ry for your smart­phone to do without the internet.”

Energy efficiency is synonymous with optimisation

There are the­re­fore seve­ral dimen­sions to ener­gy effi­cien­cy. It is not enough to reduce the num­ber of para­me­ters requi­red for model cal­cu­la­tions, whe­ther during trai­ning or infe­rence, as the Deep­Seek model has done. Action must also be taken on the trai­ning data and the data that forms the model’s know­ledge. One solu­tion that stands out is Mis­tral, an open-source French lan­guage model. There is value in divi­ding the main model into a mul­ti­tude of experts that can be acti­va­ted accor­ding to the task at hand. This is one of the approaches pro­po­sed for opti­mi­sing these models : dis­tin­gui­shing them by spe­cia­li­ty. “The goal is to take pre-trai­ned models and imple­ment stra­te­gies to adapt them with as few para­me­ters as pos­sible to dif­ferent, very spe­ci­fic sub­tasks,” explains Enzo Tar­ta­glione. “This not only avoids the impact of retrai­ning, but also great­ly improves ener­gy and performance.”

With more spe­cia­li­sed models, the amount of know­ledge requi­red to achieve this will also require less data to be acqui­red. This type of model could the­re­fore act local­ly and com­mu­ni­cate with ser­vers in a much more stream­li­ned way. After all, AI is still a rela­ti­ve­ly recent inno­va­tion. And, like most tech­no­lo­gi­cal inno­va­tions, it should logi­cal­ly fol­low the same path of opti­mi­sa­tion as its pre­de­ces­sors. Ulti­ma­te­ly, fru­gal AI is more of a move­ment in AI research than a field in its own right. In fact, com­pu­ter science has always sought to desi­gn sys­tems that opti­mise resources and limit unne­ces­sa­ry cal­cu­la­tions – fru­ga­li­ty is the­re­fore a natu­ral conti­nua­tion of this logic of effi­cien­cy. Per­haps in the same way that com­pu­ters and phones have become portable ?

In any case, the appeal of fru­ga­li­ty, in addi­tion to being envi­ron­men­tal, is also eco­no­mic for the various players deve­lo­ping this type of tool. This does mean, howe­ver, that there is still a risk of a rebound effect : more wides­pread use due to lower deve­lop­ment costs would great­ly reduce the envi­ron­men­tal bene­fits. Howe­ver, this approach will undoub­ted­ly not be the only solu­tion to the ener­gy abyss that AI repre­sents ; addres­sing the issue of sus­tai­na­bi­li­ty will also be essential…

Pablo Andres
1Pat­ter­son, D., Gon­za­lez, J., Le, Q., Liang, C., Mun­guia, L. M., Roth­child, D., … & Hen­nes­sy, J. (2021). Car­bon emis­sions and large neu­ral net­work trai­ning. arXiv pre­print arXiv:2104.10350.https://​arxiv​.org/​a​b​s​/​2​1​0​4​.​10350

Support accurate information rooted in the scientific method.

Donate