1_decisionAlgorithme
π Digital π Science and technology
What are the next challenges for AI?

“Decisions made by algorithms must be justified”

with Sophy Caulier, Independant journalist
On December 1st, 2021 |
4min reading time
Isabelle Bloch
Isabelle Bloch
Professor at Sorbonne University (Chair in Artificial Intelligence)
Key takeaways
  • Symbolic AI is based on certain rules, which reproduce human reasoning. This approach is said to be “inherently explainable”, with a few exceptions.
  • Statistical approaches to AI rely on statistical learning methods; it is difficult to extract and express the rules of what its neural networks do.
  • The need for explainability comes from different issues around trust, ethics, responsibility, and also possibly economic issues.
  • Hybrid AI can address this problem by combining several AI approaches. It combines knowledge and data, symbolic AI and neural networks, logic, and learning.
  • But, whatever the approach, the role of the human being remains essential, and it will always be necessary to justify the decisions made by an algorithm.

How and why should we explain the decisions made by arti­fi­cial intel­li­gence (AI) algorithms?

The need for explain­ab­il­ity is not new! The ques­tion was already being asked as far back as ancient times, even if back then it was from a philo­soph­ic­al point of view. It was later posed in a form­al way at the end of the 19th cen­tury, par­tic­u­larly since the work of Charles Peirce. This Amer­ic­an philo­soph­er and the­or­ist intro­duced abduct­ive reas­on­ing, i.e. the search for explan­a­tions. Many of the meth­ods used in sym­bol­ic AI, which are based on know­ledge mod­el­ling with approaches such as logic, sym­bol­ic learn­ing, etc., are said to be ‘inher­ently explain­able’, because the sequence of reas­on­ing that leads to a decision is iden­ti­fied. But this is only par­tially true, because if the prob­lem posed becomes too large, with a large num­ber of logic­al for­mu­las, very com­plex decision trees, and very numer­ous asso­ci­ation rules, explan­a­tion becomes difficult.

The ques­tion of explain­ab­il­ity is all the more rel­ev­ant today as the second paradigm of AI, stat­ist­ic­al approaches to AI, has been at the fore­front in recent years. While sym­bol­ic AI is based on rules and repro­duces human reas­on­ing, stat­ist­ic­al approaches to AI rely on stat­ist­ic­al learn­ing meth­ods, in par­tic­u­lar arti­fi­cial neur­al net­works that are trained on large volumes of data. These approaches are part of what is known as machine learn­ing (ML), includ­ing deep learn­ing (DL). It is very dif­fi­cult to extract and express the rules of what neur­al net­works do, which begin with the data.

How can an AI decision be explained?

First of all, it is neces­sary to define what to explain, for whom, how and why. The choice of explain­ab­il­ity tools or meth­ods depends on the answer to these ques­tions. For neur­al net­works, it is pos­sible to answer them at the level of the data used, at the level of the oper­a­tion of the net­work itself or at the level of the res­ult pro­duced. For the oper­a­tion, one may ask wheth­er it is neces­sary to explain. Take aspir­in for example, for a long time it was used without any­one know­ing how it worked. When it the way it worked was finally under­stood, it was used to devel­op new things, without chan­ging the way aspir­in itself was used. In the same way, you can drive a car without under­stand­ing the engine but with a level of know­ledge that is suf­fi­cient to use a car well.

At the level of the res­ult, the explan­a­tion may need to go through inter­me­di­ate steps to explain the final res­ult. For example, I work with radi­olo­gists on meas­ur­ing the thick­ness of the cor­pus cal­losum in pre­ma­ture babies. The radi­olo­gists wanted to know where the res­ults came from, which region was recog­nised in the image, where the meas­ure­ments were made, to under­stand what con­trib­uted to the decision and explain the final res­ult. These steps were neces­sary for them to have con­fid­ence in the tool.

An algorithm is expec­ted to be neut­ral, but noth­ing is ever neut­ral! The doc­tor trig­gers an ima­ging test for his patient because he is look­ing for some­thing that he can identi­fy in this image, he has an inten­tion. This intro­duces biases, which are not stat­ist­ic­al, but cog­nit­ive, of fram­ing, con­firm­a­tion, com­pla­cency, etc. These same biases are found in the face of images that have been taken. We are faced with these same biases when it comes to the res­ults pro­duced by an algorithm. Fur­ther­more, we should not for­get that we trust the algorithm more if it shows us what we are look­ing for. Anoth­er factor to con­sider is the cost of an error, which can be very dif­fer­ent depend­ing on wheth­er or not any­thing has been detec­ted. Radi­olo­gists gen­er­ally prefer to have a high­er num­ber of false pos­it­ives (since oth­er exam­in­a­tions will always con­firm or inval­id­ate what has been detec­ted) than false neg­at­ives. It is when the algorithm does not detect any­thing that it must not be mis­taken, even if the doc­tors always veri­fy the res­ults visually.

Explain­ab­il­ity there­fore var­ies accord­ing to the user and how an algorithm is used?

Explan­a­tion is a pro­cess of con­ver­sa­tion, of com­mu­nic­a­tion. We adapt the level of explan­a­tion accord­ing to the per­son we are talk­ing to. To stay with­in the med­ic­al frame­work, let’s take the example of an image show­ing a tumour. The doc­tor will explain this image and the tumour dif­fer­ently depend­ing on wheth­er he is talk­ing to his staff, to stu­dents, to an audi­ence in a con­fer­ence or to his patient. This is why doc­tors do not want the res­ults from algorithms to be made part of the patient’s records before hav­ing had a chance to check them themselves.

We also need to ask ourselves why we want to explain. Is it to jus­ti­fy, to con­trol the func­tion­ing of an algorithm, to dis­cov­er sci­entif­ic know­ledge, a phe­nomen­on? The object­ives vary and will there­fore require dif­fer­ent tools. The stakes also dif­fer, there are issues of trust, eth­ics, respons­ib­il­ity, and pos­sibly eco­nom­ic issues.

Why is the need for explic­ab­il­ity stronger at the moment?

This is mainly due to the increas­ing use of deep neur­al net­works, which have mil­lions of para­met­ers and are extremely com­plex. There is a lot of reli­ance on data in the hope that increas­ing the volumes used will help improve the res­ults. This being said, there is a lot of domain know­ledge that could be used. This is what hybrid AI pro­poses to do, com­bin­ing sev­er­al approaches to AI. It com­bines know­ledge and data, sym­bol­ic AI and neur­al net­works, logic and learn­ing. Per­son­ally, I’m a big believ­er in this. But whatever the approach, the role of the human being remains para­mount, and the decisions made by an algorithm will always have to be justified.

Support accurate information rooted in the scientific method.

Donate