Relative Feature Importance
Interpretable Machine Learning (IML) methods are used to gain insight into the relevance of a feature of interest for the performance of a model. Commonly used IML methods differ in whether they consider features of interest in isolation, e.g., Permutation Feature Importance (PFI), or in relation to all remaining feature variables, e.g., Conditional Feature Importance (CFI). As such, the perturbation mechanisms inherent to PFI and CFI represent extreme reference points. We introduce Relative Feature Importance (RFI), a generalization of PFI and CFI that allows for a more nuanced feature importance computation beyond the PFI versus CFI dichotomy. With RFI, the importance of a feature relative to any other subset of features can be assessed, including variables that were not available at training time. We derive general interpretation rules for RFI based on a detailed theoretical analysis of the implications of relative feature relevance, and demonstrate the method's usefulness on simulated examples.
Top- König, Gunnar
- Molnar, Christoph
- Bischl, Bernd
- Grosse-Wentrup, Moritz
Category |
Paper in Conference Proceedings or in Workshop Proceedings (Paper) |
Event Title |
2020 25th International Conference On Pattern Recognition |
Divisions |
Neuroinformatics |
Subjects |
Kuenstliche Intelligenz |
Event Location |
Virtual Event |
Event Type |
Conference |
Event Dates |
10.-15.01.2021 |
Date |
10 January 2021 |
Export |