A Bayesian Account of Measures of Interpretability in Human-AI Interaction

11/22/2020
by   Sarath Sreedharan, et al.
0

Existing approaches for the design of interpretable agent behavior consider different measures of interpretability in isolation. In this paper we posit that, in the design and deployment of human-aware agents in the real world, notions of interpretability are just some among many considerations; and the techniques developed in isolation lack two key properties to be useful when considered together: they need to be able to 1) deal with their mutually competing properties; and 2) an open world where the human is not just there to interpret behavior in one specific form. To this end, we consider three well-known instances of interpretable behavior studied in existing literature – namely, explicability, legibility, and predictability – and propose a revised model where all these behaviors can be meaningfully modeled together. We will highlight interesting consequences of this unified model and motivate, through results of a user study, why this revision is necessary.

READ FULL TEXT

page 3

page 6

page 7

research
04/21/2021

A Unifying Bayesian Formulation of Measures of Interpretability in Human-AI

Existing approaches for generating human-aware agent behaviors have cons...
research
06/10/2016

The Mythos of Model Interpretability

Supervised machine learning models boast remarkable predictive capabilit...
research
05/29/2018

Human-in-the-Loop Interpretability Prior

We often desire our models to be interpretable as well as accurate. Prio...
research
11/18/2016

Learning Interpretability for Visualizations using Adapted Cox Models through a User Experiment

In order to be useful, visualizations need to be interpretable. This pap...
research
02/02/2021

Evaluating the Interpretability of Generative Models by Interactive Reconstruction

For machine learning models to be most useful in numerous sociotechnical...
research
10/22/2018

Assessing the Stability of Interpretable Models

Interpretable classification models are built with the purpose of provid...
research
11/22/2022

Towards Human-Interpretable Prototypes for Visual Assessment of Image Classification Models

Explaining black-box Artificial Intelligence (AI) models is a cornerston...

Please sign up or login with your details

Forgot password? Click here to reset