Global Aggregations of Local Explanations for Black Box models

07/05/2019
by   Ilse van der Linden, et al.
0

The decision-making process of many state-of-the-art machine learning models is inherently inscrutable to the extent that it is impossible for a human to interpret the model directly: they are black box models. This has led to a call for research on explaining black box models, for which there are two main approaches. Global explanations that aim to explain a model's decision making process in general, and local explanations that aim to explain a single prediction. Since it remains challenging to establish fidelity to black box models in globally interpretable approximations, much attention is put on local explanations. However, whether local explanations are able to reliably represent the black box model and provide useful insights remains an open question. We present Global Aggregations of Local Explanations (GALE) with the objective to provide insights in a model's global decision making process. Overall, our results reveal that the choice of aggregation matters. We find that the global importance introduced by Local Interpretable Model-agnostic Explanations (LIME) does not reliably represent the model's global behavior. Our proposed aggregations are better able to represent how features affect the model's predictions, and to provide global insights by identifying distinguishing features.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/26/2018

Please Stop Explaining Black Box Models for High Stakes Decisions

There are black box models now being used for high stakes decision-makin...
research
07/14/2023

Dissenting Explanations: Leveraging Disagreement to Reduce Model Overreliance

While explainability is a desirable characteristic of increasingly compl...
research
06/22/2017

MAGIX: Model Agnostic Globally Interpretable Explanations

Explaining the behavior of a black box machine learning model at the ins...
research
05/18/2023

BELLA: Black box model Explanations by Local Linear Approximations

In recent years, understanding the decision-making process of black-box ...
research
12/23/2021

AcME – Accelerated Model-agnostic Explanations: Fast Whitening of the Machine-Learning Black Box

In the context of human-in-the-loop Machine Learning applications, like ...
research
01/28/2022

Locally Invariant Explanations: Towards Stable and Unidirectional Explanations through Local Invariant Learning

Locally interpretable model agnostic explanations (LIME) method is one o...
research
06/21/2023

Investigating Poor Performance Regions of Black Boxes: LIME-based Exploration in Sepsis Detection

Interpreting machine learning models remains a challenge, hindering thei...

Please sign up or login with your details

Forgot password? Click here to reset