Pitfalls to Avoid when Interpreting Machine Learning Models

Pitfalls to Avoid when Interpreting Machine Learning Models

Abstract

Modern requirements for machine learning (ML) models include both high predictive performance and model interpretability. A growing number of techniques provide model interpretations, but can lead to wrong conclusions if applied incorrectly. We illustrate pitfalls of ML model interpretation such as bad model generalization, dependent features, feature interactions or unjustified causal interpretations. Our paper addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research.

Grafik Top
Authors
  • Molnar, Christoph
  • König, Gunnar
  • Herbinger, Julia
  • Freiesleben, Timo
  • Dandl, Susanne
  • Scholbeck, Christian A.
  • Casalicchio, Giuseppe
  • Grosse-Wentrup, Moritz
  • Bischl, Bernd
Grafik Top
Shortfacts
Category
Paper in Conference Proceedings or in Workshop Proceedings (Paper)
Event Title
XXAI: Extending Explainable AI Beyond Deep Models and Classifiers, ICML 2020 Workshop
Divisions
Neuroinformatics
Event Location
Virtually
Event Type
Workshop
Event Dates
18th of July, 2020
Date
18 July 2020
Export
Grafik Top