Back to Search
Start Over
Quod erat demonstrandum? - Towards a typology of the concept of explanation for the design of explainable AI.
- Source :
-
Expert Systems with Applications . Mar2023:Part A, Vol. 213, pN.PAG-N.PAG. 1p. - Publication Year :
- 2023
-
Abstract
- In this paper, we present a fundamental framework for defining different types of explanations of AI systems and the criteria for evaluating their quality. Starting from a structural view of how explanations can be constructed, i.e., in terms of an explanandum (what needs to be explained), multiple explanantia (explanations, clues, or parts of information that explain), and a relationship linking explanandum and explanantia, we propose an explanandum-based typology and point to other possible typologies based on how explanantia are presented and how they relate to explanandia. We also highlight two broad and complementary perspectives for defining possible quality criteria for assessing explainability: epistemological and psychological (cognitive). These definition attempts aim to support the three main functions that we believe should attract the interest and further research of XAI scholars: clear inventories, clear verification criteria, and clear validation methods. • We propose a framework for defining different types of explanations of AI systems. • We contextualize current XAI discourses within the proposed framework. • We highlight two broad perspectives for defining quality criteria for explainability. • We discuss the relevance of our framework in light of current and upcoming AI regulation. • We confer fundamental aspects for future research of XAI scholars. [ABSTRACT FROM AUTHOR]
- Subjects :
- *ARTIFICIAL intelligence
*EXPLANATION
*MACHINE learning
Subjects
Details
- Language :
- English
- ISSN :
- 09574174
- Volume :
- 213
- Database :
- Academic Search Index
- Journal :
- Expert Systems with Applications
- Publication Type :
- Academic Journal
- Accession number :
- 160292407
- Full Text :
- https://doi.org/10.1016/j.eswa.2022.118888