Skip to Main content Skip to Navigation
Conference papers

Identifying the "Right" Level of Explanation in a Given Situation

Abstract : We present a framework for defining the "right" level of AI explainability based on technical, legal and economic considerations. Our approach involves three logical steps: First, define the main con-textual factors, such as who is the audience of the explanation, the operational context, the level of harm that the system could cause, and the legal/regulatory framework. This step will help characterize the operational and legal needs for explanation, and the corresponding social benefits. Second, examine the technical tools available, including post-hoc approaches (input perturbation, saliency maps...) and hybrid AI approaches. Third, as function of the first two steps, choose the right levels of global and local explanation outputs, taking into the account the costs involved. We identify seven kinds of costs and emphasize that explanations are socially useful only when total social benefits exceed costs.
Complete list of metadata

Cited literature [26 references]  Display  Hide  Download
Contributor : David Bounie Connect in order to contact the contributor
Submitted on : Friday, March 13, 2020 - 8:11:29 AM
Last modification on : Wednesday, May 4, 2022 - 12:19:14 PM
Long-term archiving on: : Sunday, June 14, 2020 - 12:51:11 PM


Files produced by the author(s)


  • HAL Id : hal-02507316, version 1



Valérie Beaudouin, Isabelle Bloch, David Bounie, Stéphan Clémençon, Florence d'Alché-Buc, et al.. Identifying the "Right" Level of Explanation in a Given Situation. First International Workshop on New Foundations for Human-Centered AI (NeHuAI), Sep 2020, Santiago de Compostella, Spain. pp.63. ⟨hal-02507316⟩



Record views


Files downloads