Abstract

With the broader and highly successful usage of machine learning (ML) in industry and the sciences, there has been a growing demand for explainable artificial intelligence (XAI). Interpretability and explanation methods for gaining a better understanding of the problem-solving abilities and strategies of nonlinear ML, in particular, deep neural networks, are, therefore, receiving increased attention. In this work, we aim to: 1) provide a timely overview of this active emerging field, with a focus on 'post hoc' explanations, and explain its theoretical foundations; 2) put interpretability algorithms to a test both from a theory and comparative evaluation perspective using extensive simulations; 3) outline best practice aspects, i.e., how to best include interpretation methods into the standard usage of ML; and 4) demonstrate successful usage of XAI in a representative selection of application scenarios. Finally, we discuss challenges and possible future directions of this exciting foundational field of ML.

Keywords

InterpretabilityArtificial intelligenceComputer scienceMachine learningField (mathematics)Perspective (graphical)Artificial neural networkInterpretation (philosophy)Deep neural networksSelection (genetic algorithm)Deep learningManagement scienceData scienceEngineering

Affiliated Institutions

Related Publications

Explainable Artificial Intelligence (XAI)

Complex machine learning models perform better. However, we consider these models as black boxes. That’s where Explainable AI (XAI) comes into play. Understanding why a model ma...

2021 Zenodo (CERN European Organization fo... 543 citations

Publication Info

Year
2021
Type
review
Volume
109
Issue
3
Pages
247-278
Citations
1115
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1115
OpenAlex

Cite This

Klaus‐Robert Müller, Wojciech Samek, Grégoire Montavon et al. (2021). Explaining Deep Neural Networks and Beyond: A Review of Methods and Applications. Proceedings of the IEEE , 109 (3) , 247-278. https://doi.org/10.1109/jproc.2021.3060483

Identifiers

DOI
10.1109/jproc.2021.3060483