1_Comprendre
π Science and technology π Digital
Generative AI: threat or opportunity?

ChatGPT, Midjourney : everything you need to know about generative AI

with Éric Moulines, Professor of Statistical Learning at École Polytechnique (IP Paris), Hatim Bourfoune, AI research engineer at IDRIS (CNRS) and Pierre Cornette, AI support engineer at IDRIS (CNRS)
On November 21st, 2023 |
5 min reading time
Eric Moulines
Éric Moulines
Professor of Statistical Learning at École Polytechnique (IP Paris)
Hatim Bourfoune
Hatim Bourfoune
AI research engineer at IDRIS (CNRS)
Pierre Cornette
Pierre Cornette
AI support engineer at IDRIS (CNRS)
Key takeaways
  • Generative AI can create content from a database that has been ingested and according to the indications they are given.
  • These technologies, which remain new, are still being developed and there are still several areas for improvement: reliability, bias in the database, etc.
  • ChatGPT and Bloom are just two models of generative AI, but the concept can be extended to a multitude of applications.
  • These technologies raise a few questions, such as their ecological impact and the risk of using them for potentially malicious purposes.

It’s all the talk these days, and ChatGPT is arri­ving in our socie­ties like a veri­table revo­lu­tion. So, it is hard­ly sur­pri­sing that, given the wide-ran­ging appli­ca­tions of these tools, their arri­val is fuel­ling so much debate. But do we real­ly know how this AI works ?

A gene­ra­tive AI can gene­rate writ­ten, visual, or audible content by inges­ting content. By giving it indi­ca­tions as input, the AI can create as out­put any content that cor­res­ponds to the indi­ca­tions inges­ted. “Here, we’re loo­king to gene­rate ori­gi­nal content,” explains Éric Mou­lines, pro­fes­sor of sta­tic lear­ning at École Poly­tech­nique (IP Paris). “This ori­gi­nal content will be obtai­ned by gene­ra­li­sing the infor­ma­tion seen during learning”.

There are cur­rent­ly two main types of AI model. GPTs (Gene­ra­tive Pre-trai­ned Trans­for­mers), such as ChatGPT, and Dif­fu­sion Models. “By giving it text as input, the AI will be able to unders­tand the context through a mecha­nism cal­led atten­tion,” adds Hatim Bour­fone, AI research engi­neer at IDRIS (CNRS). “Its out­put will the­re­fore be a list of all the words in the dic­tio­na­ry that it knows [lear­ned during trai­ning phase] on which it will have pla­ced a pro­ba­bi­li­ty”. Depen­ding on the data­base it has trai­ned on, the tool can be pro­gram­med for various functions.

Bloom, for example, the AI deve­lo­ped by the team of which Hatim Bour­foune is a mem­ber at IDRIS, is a tool that helps resear­chers to express them­selves in seve­ral lan­guages. “The pri­ma­ry aim of the Bloom model,” adds Pierre Cor­nette, also a mem­ber of the IDRIS team, “is to learn a lan­guage. To do this, we give them a whole bunch of texts to ingest, asking them to pre­dict the next word in the given text, and we cor­rect them if they get it wrong”. 

A recent, still immature technology

“The first gene­ra­tive AI models are not even 10 years old,” explains Éric Mou­lines. “The first revo­lu­tion in this field was the arri­val of trans­for­mers – a tech­no­lo­gy per­fec­ting this atten­tion mecha­nism – in 2017. Four years later, we alrea­dy have com­mer­cial pro­ducts. So, there has been consi­de­rable acce­le­ra­tion, much fas­ter than on any other Deep Lear­ning model.” Models like ChatGPT are the­re­fore still very new, and there are still many things that can, or must, be improved.

The ques­tion of the relia­bi­li­ty of the ans­wers given is still not cer­tain : “ChatGPT is not fami­liar with the notion of relia­bi­li­ty”, admits the pro­fes­sor. “This type of AI is inca­pable of asses­sing the vera­ci­ty of the ans­wers it gives.” This leaves room for an easi­ly obser­vable phe­no­me­non known as ‘hal­lu­ci­na­tions’. “It is pos­sible [for ChatGPT] to gene­rate content that seems plau­sible, but is rigo­rous­ly false,” he adds. “It uses com­ple­te­ly pro­ba­bi­lis­tic rea­so­ning to gene­rate sequences of words. Depen­ding on the context, it will gene­rate strings of words that seem the most likely.”

Apart from its abi­li­ty to invent book titles, other limi­ta­tions should be borne in mind when using it. By applying Deep Lear­ning methods, these AIs go through a trai­ning phase during which they ingest a quan­ti­ty of exis­ting texts. In this way, they will incor­po­rate the biases of this data­base into their lear­ning. Geo­po­li­ti­cal ques­tions are a good example of this. “If you ask it geo­po­li­ti­cal ques­tions, ChatGPT will essen­tial­ly reflect the Wes­tern world,” says Éric Mou­lines. “If we show the ans­wers given to a Chi­nese per­son, he will cer­tain­ly not agree with what is said about the sove­rei­gn­ty of such and such a coun­try over a given territory.”

A range of applications

Each model will the­re­fore be able to gene­rate content accor­ding to the data­base it has been trai­ned on. This is per­haps where the magic of this tech­no­lo­gy lies, because, kno­wing this, a myriad of appli­ca­tions can be crea­ted. “A good ana­lo­gy for this tech­no­lo­gy would be that of an engine,” says Pierre Cor­nette. “You could have a very power­ful engine, but it can be used for either a trac­tor or a racing car.” For example, ChatGPT is a racing car, and its engine is GPT‑4. “The advan­tage is that the tech­no­lo­gies are concen­tra­ted in what is the engine,” he conti­nues, “and you don’t need to unders­tand how it works to use the race car.”

Bloom is an example of ano­ther use for this type of model : “A year ago, Bloom was one of the only models that was com­ple­te­ly open to research,” insists Hatim Bour­foune. In other words, anyone could down­load the model and use it for their own research. Trai­ned with a data­base of various scien­ti­fic articles in many lan­guages, this model can be extre­me­ly use­ful for scien­ti­fic research. Pierre Cor­nette adds : “There is also ano­ther Big­code pro­ject, run by the same people, which pro­motes a model spe­cia­li­sing in com­pu­ter code. We ask it for a func­tion, sim­ply des­cri­bing its action, and it can write it for us in the desi­red language.”

The popu­la­ri­ty of ChatGPT shows just how impor­tant it is for the gene­ral public. Bing has also inte­gra­ted it into its search engine with a view to com­pe­ting with Google. This inte­gra­tion makes it pos­sible to coun­ter one of the limi­ta­tions of this tech­no­lo­gy : the relia­bi­li­ty of the ans­wers given. By giving the sources used to com­pile its res­ponse, the search engine enables us to unders­tand and veri­fy them bet­ter. Even more recent­ly, Adobe has inte­gra­ted a gene­ra­tive AI model into various soft­ware appli­ca­tions in its suite (such as Pho­to­shop and Illus­tra­tor), revea­ling yet ano­ther impres­sive appli­ca­tion of this technology.

“An exciting future”

All this can only mean an exci­ting future for this inno­va­tion. Howe­ver, the range of appli­ca­tions raises ques­tions about its pos­sible uses. “As with all tools, there can be mali­cious uses,” admits Hatim Bour­foune. “That’s why com­pa­nies like Ope­nAI put up dif­ferent secu­ri­ty bar­riers.” Today, many of the ques­tions put to ChatGPT remain unans­we­red, because the AI believes that they vio­late its content policy.

Even so, this tech­no­lo­gy is still in its infan­cy. “That’s the prin­ciple of research – we’re still at ground zero,” says Éric Mou­lines. “It’s ama­zing that it even works.” There are still many loo­pholes to be filled, par­ti­cu­lar­ly from a legal point of view. As explai­ned, the content gene­ra­ted by these tools will be built using an exis­ting data­base. The AI will the­re­fore “copy” exis­ting texts or works without citing their ori­gi­nal author. “This poses a major pro­blem,” he conti­nues, “because the rights hol­ders of the content used to gene­rate these new images [or texts] are not respected.”

Des­pite its various limi­ta­tions, the poten­tial remains enor­mous : “What excites me … is that the pro­gress to be made is enor­mous,” adds the pro­fes­sor. “But the trend and the deri­va­tives are enor­mous. It’s hap­pe­ning very qui­ck­ly and the­re’s very exci­ting com­pe­ti­tion in these sub­jects.” Spea­king of deri­va­tives, Bloom illus­trates this per­fect­ly. Use­ful for research, it is also a lin­guis­tic tool that could make it pos­sible to save dead lan­guages, but also to trans­late scien­ti­fic texts into les­ser-spo­ken lan­guages to faci­li­tate the dis­se­mi­na­tion of research.

Howe­ver, its “exci­ting” future may be ham­pe­red by its consi­de­rable car­bon impact. “These models require a lot of memo­ry, because they need to store a huge amount of data,” explains Éric Mou­lines. “Today, we esti­mate that Ope­nAI consumes as much memo­ry as the grid in a coun­try like Bel­gium.” This is the pro­blem that will sur­ely be the most com­pli­ca­ted to solve.

Pablo Andres

Support accurate information rooted in the scientific method.

Donate