Pitfalls to Avoid when Interpreting Machine Learning Models
Abstract
Modern requirements for machine learning (ML) models include both high predictive performance and model interpretability. A growing number of techniques provide model interpretations, but can lead to wrong conclusions if applied incorrectly. We illustrate pitfalls of ML model interpretation such as bad model generalization, dependent features, feature interactions or unjustified causal interpretations. Our paper addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research.
Top- Molnar, Christoph
- König, Gunnar
- Herbinger, Julia
- Freiesleben, Timo
- Dandl, Susanne
- Scholbeck, Christian A.
- Casalicchio, Giuseppe
- Grosse-Wentrup, Moritz
- Bischl, Bernd
Shortfacts
Category |
Paper in Conference Proceedings or in Workshop Proceedings (Paper) |
Event Title |
XXAI: Extending Explainable AI Beyond Deep Models and Classifiers, ICML 2020 Workshop |
Divisions |
Neuroinformatics |
Event Location |
Virtually |
Event Type |
Workshop |
Event Dates |
18th of July, 2020 |
Date |
18 July 2020 |
Export |