neuroscienceEtSociete_05curiosite
π Society π Neuroscience
Neuroscience: our relationship with intelligence

Mind of an AI : from curiosity to autonomy

par Agnès Vernet, Science journalist
On February 18th, 2021 |
4min reading time
Pierre-Yves Oudeyer
Pierre-Yves Oudeyer
Inria Research director and head of FLOWERS team at Inria/Ensta Paris (IP Paris)
Key takeaways
  • Research in artificial intelligence feeds off cognitive sciences, but now neurobiology is also making progress thanks to algorithmic models.
  • Curiosity, referred to as “intrinsic motivation” by psychologists, is a necessary trait for independent learning in children.
  • For Dr. Pierre-Yves Oudeyer, this mechanism can also be applied to machines.
  • As such, his research explores human cognition to improve artificial intelligence… and vice-versa.

How does one mea­sure the intel­li­gence of an arti­fi­cial intelligence ? 

It isn’t easy because the term “arti­fi­cial intel­li­gence” is used by the gene­ral public to refer to objects deve­lo­ped in this research field, such as soft­ware equip­ped with a lear­ning sys­tem. In fact, arti­fi­cial intel­li­gence is not a thing. Rather it is a field of stu­dy which tries to model func­tions of the human mind, like memo­ry, rea­so­ning, lear­ning or lan­guage. We the­re­fore can­not mea­sure intelligence.

Moreo­ver, the notion of ‘intel­li­gence’ makes no sense gene­ral­ly spea­king. For example, it’s impos­sible to say that an ear­th­worm is more stu­pid than a human. Ins­tead, each living being has beha­vio­ral and mor­pho­lo­gi­cal cha­rac­te­ris­tics resul­ting from an evo­lu­tio­na­ry pro­cess rela­ted to its par­ti­cu­lar envi­ron­ment. Ear­th­worms can find food in the ground. And, in their own eco­sys­tem, human beings have social inter­ac­tions – lin­guis­tic or cultu­ral – with others. An ear­th­worm wouldn’t know what to do in our eco­sys­tem and a human wouldn’t do any bet­ter in soil.

Tech­no­lo­gies, too, are deve­lo­ped in very spe­ci­fic contexts. It can­not be said that smart­phone voice recog­ni­tion sys­tems are stu­pid because they don’t unders­tand the mea­ning of the sen­tences that they trans­cribe. They were not trai­ned to that end, it’s not part of their ‘eco­sys­tem’. 

The soft­ware you men­tion trans­cribes and learns, could it also understand ?

Fun­da­men­tal­ly, the mea­ning we asso­ciate with a sen­tence is per­so­ni­fied, it is inter­pre­ted based on the sen­so­ry and motor expe­riences of our body in its envi­ron­ment. If a machine does not have access to a body to phy­si­cal­ly inter­act with our world, it stands no chance of inter­pre­ting sen­tences like we do.

Howe­ver, we can train lan­guage models with large text data­bases. Machines can then detect sta­tis­ti­cal pat­terns and do asto­ni­shing tasks, such as ans­we­ring a simple ques­tion, by pre­dic­ting the struc­tures of sen­tences accor­ding to a given context. These tools are very use­ful in the world of indus­try, for human-machine inter­face, when machines must inter­pret direc­tives based on the context. In order to do this, unlike humans, they don’t neces­sa­ri­ly need to unders­tand sentences.

In your research you state that, in humans, part of lear­ning is dri­ven by curio­si­ty. Can it be applied to software ? 

This issue is at the heart of my team’s research. We stu­dy the mecha­nisms of curio­si­ty, or what psy­cho­lo­gists call “intrin­sic moti­va­tion”. It allows living beings to under­take inde­pendent lear­ning. We deve­lop algo­rith­mic models of curio­si­ty to high­light the mecha­nisms invol­ved, such as spon­ta­neous explo­ra­tion. This plays a fun­da­men­tal role in the sen­so­ry, cog­ni­tive and motor deve­lop­ment of humans.

We then test our theo­ries with volun­teers or machines. In doing so, we dis­co­ve­red that to effi­cient­ly explore its envi­ron­ment, a robot must pick the areas where it makes the most pro­gress, mea­ning those in which the gap bet­ween pre­dic­tion and rea­li­ty tends to decrease. For example, it is in its inter­est to play with an object with which he tends to make pro­gress rather than ano­ther that he imme­dia­te­ly mas­ters or, on the contra­ry, that he can­not use at all. We sho­wed that, in theo­ry, this stra­te­gy is effi­cient for robots. The ques­tion remains open as to whe­ther humans use this mea­sure to make pro­gress and to guide the explo­ra­tion of their surroundings.

But could this mea­sure of pro­gress explain why humans tend to pre­fer acti­vi­ties that they are able to learn easily ?

Yes. The mecha­nism of explo­ra­tion accor­ding to pro­gress leads to a “snow­ball” effect : when explo­ring an acti­vi­ty, ini­tia­ted ran­dom­ly or other contin­gent fac­tors, we deve­lop know­ledge or skills which will make simi­lar types of acti­vi­ties easier to learn. This encou­rages the indi­vi­dual to pur­sue that course of action ; also asso­cia­ted with the plea­sure res­ponse in the brain upon explo­ring new activities.

This fun­da­men­tal hypo­the­sis explains the diver­si­ty of lear­ning tra­jec­to­ries found in dif­ferent people. To confirm that, we com­pa­red the beha­vior of adult volun­teers to that pre­dic­ted by our com­pu­ta­tio­nal model. These ana­lyses sho­wed that lear­ning pro­gress and per­for­mance for each task are mea­sures used by humans to guide their explo­ra­tion. Both work in dif­ferent ways : the com­bi­na­tion of these dif­fe­rences and the snow­ball effect men­tio­ned above sup­ports the idea that there is a diver­si­ty in lear­ning path­ways, which explains dif­fe­rences bet­ween individuals.

Does this model improve machines ?

Our theo­ries can some­times be imple­men­ted in machines to make them more adap­table. But the explo­ra­to­ry beha­vior of humans is not neces­sa­ri­ly the most opti­mal choice. For example, other curio­si­ty mecha­nisms are bet­ter sui­ted for robots des­ti­ned to auto­no­mous­ly explore the ocean floor or the sur­face of Mars – if only to prevent these machines from making dan­ge­rous choices as much as possible.

Can these tools also help humans to learn better ?

Yes, there are indeed appli­ca­tions in the field of edu­ca­tion. We desi­gned soft­ware that crea­ted cus­tom exer­cise sequences for stu­dents in mathe­ma­tics. The objec­tive is to adapt a series to each child that will opti­mise both his/her lear­ning and moti­va­tion. We know that the lat­ter is an impor­tant fac­tor in aca­de­mic fai­lure. Moti­va­tio­nal aspects prompt stu­dents to per­se­vere and try har­der. Using curio­si­ty models, we deve­lo­ped algo­rithms that inter­act with each child indi­vi­dual­ly to offer a moti­va­tio­nal series of exer­cises accor­ding to the child’s profile.

In a pre­vious pro­ject named Kid­learn we sho­wed that, on ave­rage, a grea­ter diver­si­ty of stu­dents made more pro­gress thanks to our software’s pro­po­sals than with a tea­ching expert. A figure that includes chil­dren with a range of dif­fi­cul­ties or abi­li­ties, too. This bene­fit was asso­cia­ted with a higher degree of intrin­sic moti­va­tion. We are now wor­king with a consor­tium of indus­trials in the field of edu­ca­tion tech­no­lo­gy (edTech) in order to trans­fer this approach into a digi­tal edu­ca­tio­nal sys­tem inten­ded to be used on a large scale in pri­ma­ry schools in France (Adap­tiv’­Maths pro­ject). My col­league Hélène Sau­zéon even sho­wed that this sys­tem can faci­li­tate lear­ning for chil­dren suf­fe­ring from deve­lop­men­tal disor­ders such as autism.

Support accurate information rooted in the scientific method.

Donate