Implicit Mixture of Interpretable Experts for Global and Local Interpretability

12/01/2022
by   Nathan Elazar, et al.
0

We investigate the feasibility of using mixtures of interpretable experts (MoIE) to build interpretable image classifiers on MNIST10. MoIE uses a black-box router to assign each input to one of many inherently interpretable experts, thereby providing insight into why a particular classification decision was made. We find that a naively trained MoIE will learn to 'cheat', whereby the black-box router will solve the classification problem by itself, with each expert simply learning a constant function for one particular class. We propose to solve this problem by introducing interpretable routers and training the black-box router's decisions to match the interpretable router. In addition, we propose a novel implicit parameterization scheme that allows us to build mixtures of arbitrary numbers of experts, allowing us to study how classification performance, local and global interpretability vary as the number of experts is increased. Our new model, dubbed Implicit Mixture of Interpretable Experts (IMoIE) can match state-of-the-art classification accuracy on MNIST10 while providing local interpretability, and can provide global interpretability albeit at the cost of reduced classification accuracy.

READ FULL TEXT

page 9

page 10

research
02/10/2020

Interpretable Companions for Black-Box Models

We present an interpretable companion model for any pre-trained black-bo...
research
06/05/2022

Interpretable Mixture of Experts for Structured Data

With the growth of machine learning for structured data, the need for re...
research
11/09/2022

Mapping the Ictal-Interictal-Injury Continuum Using Interpretable Machine Learning

IMPORTANCE: An interpretable machine learning model can provide faithful...
research
05/04/2020

Construction and Elicitation of a Black Box Model in the Game of Bridge

We address the problem of building a decision model for a specific biddi...
research
03/13/2023

Revisiting model self-interpretability in a decision-theoretic way for binary medical image classification

Interpretability is highly desired for deep neural network-based classif...
research
03/23/2021

IAIA-BL: A Case-based Interpretable Deep Learning Model for Classification of Mass Lesions in Digital Mammography

Interpretability in machine learning models is important in high-stakes ...
research
06/01/2021

Memory Wrap: a Data-Efficient and Interpretable Extension to Image Classification Models

Due to their black-box and data-hungry nature, deep learning techniques ...

Please sign up or login with your details

Forgot password? Click here to reset