1_decisionAlgorithme
π Digital π Science and technology
What are the next challenges for AI?

“Decisions made by algorithms must be justified”

with Sophy Caulier, Independant journalist
On December 1st, 2021 |
4min reading time
Isabelle Bloch
Isabelle Bloch
Professor at Sorbonne University (Chair in Artificial Intelligence)
Key takeaways
  • Symbolic AI is based on certain rules, which reproduce human reasoning. This approach is said to be “inherently explainable”, with a few exceptions.
  • Statistical approaches to AI rely on statistical learning methods; it is difficult to extract and express the rules of what its neural networks do.
  • The need for explainability comes from different issues around trust, ethics, responsibility, and also possibly economic issues.
  • Hybrid AI can address this problem by combining several AI approaches. It combines knowledge and data, symbolic AI and neural networks, logic, and learning.
  • But, whatever the approach, the role of the human being remains essential, and it will always be necessary to justify the decisions made by an algorithm.

How and why should we explain the deci­sions made by arti­fi­cial intel­li­gence (AI) algorithms ?

The need for explai­na­bi­li­ty is not new ! The ques­tion was alrea­dy being asked as far back as ancient times, even if back then it was from a phi­lo­so­phi­cal point of view. It was later posed in a for­mal way at the end of the 19th cen­tu­ry, par­ti­cu­lar­ly since the work of Charles Peirce. This Ame­ri­can phi­lo­so­pher and theo­rist intro­du­ced abduc­tive rea­so­ning, i.e. the search for expla­na­tions. Many of the methods used in sym­bo­lic AI, which are based on know­ledge model­ling with approaches such as logic, sym­bo­lic lear­ning, etc., are said to be ‘inhe­rent­ly explai­nable’, because the sequence of rea­so­ning that leads to a deci­sion is iden­ti­fied. But this is only par­tial­ly true, because if the pro­blem posed becomes too large, with a large num­ber of logi­cal for­mu­las, very com­plex deci­sion trees, and very nume­rous asso­cia­tion rules, expla­na­tion becomes difficult.

The ques­tion of explai­na­bi­li­ty is all the more rele­vant today as the second para­digm of AI, sta­tis­ti­cal approaches to AI, has been at the fore­front in recent years. While sym­bo­lic AI is based on rules and repro­duces human rea­so­ning, sta­tis­ti­cal approaches to AI rely on sta­tis­ti­cal lear­ning methods, in par­ti­cu­lar arti­fi­cial neu­ral net­works that are trai­ned on large volumes of data. These approaches are part of what is known as machine lear­ning (ML), inclu­ding deep lear­ning (DL). It is very dif­fi­cult to extract and express the rules of what neu­ral net­works do, which begin with the data.

How can an AI deci­sion be explained ?

First of all, it is neces­sa­ry to define what to explain, for whom, how and why. The choice of explai­na­bi­li­ty tools or methods depends on the ans­wer to these ques­tions. For neu­ral net­works, it is pos­sible to ans­wer them at the level of the data used, at the level of the ope­ra­tion of the net­work itself or at the level of the result pro­du­ced. For the ope­ra­tion, one may ask whe­ther it is neces­sa­ry to explain. Take aspi­rin for example, for a long time it was used without anyone kno­wing how it wor­ked. When it the way it wor­ked was final­ly unders­tood, it was used to deve­lop new things, without chan­ging the way aspi­rin itself was used. In the same way, you can drive a car without unders­tan­ding the engine but with a level of know­ledge that is suf­fi­cient to use a car well.

At the level of the result, the expla­na­tion may need to go through inter­me­diate steps to explain the final result. For example, I work with radio­lo­gists on mea­su­ring the thi­ck­ness of the cor­pus cal­lo­sum in pre­ma­ture babies. The radio­lo­gists wan­ted to know where the results came from, which region was reco­gni­sed in the image, where the mea­su­re­ments were made, to unders­tand what contri­bu­ted to the deci­sion and explain the final result. These steps were neces­sa­ry for them to have confi­dence in the tool.

An algo­rithm is expec­ted to be neu­tral, but nothing is ever neu­tral ! The doc­tor trig­gers an ima­ging test for his patient because he is loo­king for some­thing that he can iden­ti­fy in this image, he has an inten­tion. This intro­duces biases, which are not sta­tis­ti­cal, but cog­ni­tive, of fra­ming, confir­ma­tion, com­pla­cen­cy, etc. These same biases are found in the face of images that have been taken. We are faced with these same biases when it comes to the results pro­du­ced by an algo­rithm. Fur­ther­more, we should not for­get that we trust the algo­rithm more if it shows us what we are loo­king for. Ano­ther fac­tor to consi­der is the cost of an error, which can be very dif­ferent depen­ding on whe­ther or not any­thing has been detec­ted. Radio­lo­gists gene­ral­ly pre­fer to have a higher num­ber of false posi­tives (since other exa­mi­na­tions will always confirm or inva­li­date what has been detec­ted) than false nega­tives. It is when the algo­rithm does not detect any­thing that it must not be mis­ta­ken, even if the doc­tors always veri­fy the results visually.

Explai­na­bi­li­ty the­re­fore varies accor­ding to the user and how an algo­rithm is used ?

Expla­na­tion is a pro­cess of conver­sa­tion, of com­mu­ni­ca­tion. We adapt the level of expla­na­tion accor­ding to the per­son we are tal­king to. To stay within the medi­cal fra­me­work, let’s take the example of an image sho­wing a tumour. The doc­tor will explain this image and the tumour dif­fe­rent­ly depen­ding on whe­ther he is tal­king to his staff, to stu­dents, to an audience in a confe­rence or to his patient. This is why doc­tors do not want the results from algo­rithms to be made part of the patient’s records before having had a chance to check them themselves.

We also need to ask our­selves why we want to explain. Is it to jus­ti­fy, to control the func­tio­ning of an algo­rithm, to dis­co­ver scien­ti­fic know­ledge, a phe­no­me­non ? The objec­tives vary and will the­re­fore require dif­ferent tools. The stakes also dif­fer, there are issues of trust, ethics, res­pon­si­bi­li­ty, and pos­si­bly eco­no­mic issues.

Why is the need for expli­ca­bi­li­ty stron­ger at the moment ?

This is main­ly due to the increa­sing use of deep neu­ral net­works, which have mil­lions of para­me­ters and are extre­me­ly com­plex. There is a lot of reliance on data in the hope that increa­sing the volumes used will help improve the results. This being said, there is a lot of domain know­ledge that could be used. This is what hybrid AI pro­poses to do, com­bi­ning seve­ral approaches to AI. It com­bines know­ledge and data, sym­bo­lic AI and neu­ral net­works, logic and lear­ning. Per­so­nal­ly, I’m a big belie­ver in this. But wha­te­ver the approach, the role of the human being remains para­mount, and the deci­sions made by an algo­rithm will always have to be justified.

Support accurate information rooted in the scientific method.

Donate