The Solvability of Interpretability Evaluation Metrics

05/18/2022
by   Yilun Zhou, et al.
0

Feature attribution methods are popular for explaining neural network predictions, and they are often evaluated on metrics such as comprehensiveness and sufficiency, which are motivated by the principle that more important features – as judged by the explanation – should have larger impacts on model prediction. In this paper, we highlight an intriguing property of these metrics: their solvability. Concretely, we can define the problem of optimizing an explanation for a metric and solve it using beam search. This brings up the obvious question: given such solvability, why do we still develop other explainers and then evaluate them on the metric? We present a series of investigations showing that this beam search explainer is generally comparable or favorable to current choices such as LIME and SHAP, suggest rethinking the goals of model interpretability, and identify several directions towards better evaluations of new method proposals.

READ FULL TEXT
research
11/23/2022

Evaluating Feature Attribution Methods for Electrocardiogram

The performance of cardiac arrhythmia detection with electrocardiograms(...
research
06/08/2020

Evaluation Criteria for Instance-based Explanation

Explaining predictions made by complex machine learning models helps use...
research
09/18/2020

Evaluation of Local Explanation Methods for Multivariate Time Series Forecasting

Being able to interpret a machine learning model is a crucial task in ma...
research
02/11/2022

InterpretTime: a new approach for the systematic evaluation of neural-network interpretability in time series classification

We present a novel approach to evaluate the performance of interpretabil...
research
04/13/2023

Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance

Interpretability methods are valuable only if their explanations faithfu...
research
04/01/2021

Explaining COVID-19 and Thoracic Pathology Model Predictions by Identifying Informative Input Features

Neural networks have demonstrated remarkable performance in classificati...
research
11/18/2018

Regularized adversarial examples for model interpretability

As machine learning algorithms continue to improve, there is an increasi...

Please sign up or login with your details

Forgot password? Click here to reset