NxPlain: Web-based Tool for Discovery of Latent Concepts

03/06/2023
by   Fahim Dalvi, et al.
0

The proliferation of deep neural networks in various domains has seen an increased need for the interpretability of these models, especially in scenarios where fairness and trust are as important as model performance. A lot of independent work is being carried out to: i) analyze what linguistic and non-linguistic knowledge is learned within these models, and ii) highlight the salient parts of the input. We present NxPlain, a web application that provides an explanation of a model's prediction using latent concepts. NxPlain discovers latent concepts learned in a deep NLP model, provides an interpretation of the knowledge learned in the model, and explains its predictions based on the used concepts. The application allows users to browse through the latent concepts in an intuitive order, letting them efficiently scan through the most salient concepts with a global corpus level view and a local sentence-level view. Our tool is useful for debugging, unraveling model bias, and for highlighting spurious correlations in a model. A hosted demo is available here: https://nxplain.qcri.org.

READ FULL TEXT

page 4

page 5

page 6

research
11/12/2022

ConceptX: A Framework for Latent Concept Analysis

The opacity of deep neural networks remains a challenge in deploying sol...
research
05/15/2022

Discovering Latent Concepts Learned in BERT

A large number of studies that analyze deep neural network models and th...
research
02/02/2018

Causal Learning and Explanation of Deep Neural Networks via Autoencoded Activations

Deep neural networks are complex and opaque. As they enter application i...
research
10/19/2022

Towards Procedural Fairness: Uncovering Biases in How a Toxic Language Classifier Uses Sentiment Information

Previous works on the fairness of toxic language classifiers compare the...
research
07/31/2021

A Hypothesis for the Aesthetic Appreciation in Neural Networks

This paper proposes a hypothesis for the aesthetic appreciation that aes...
research
08/08/2021

Human-in-the-loop Extraction of Interpretable Concepts in Deep Learning Models

The interpretation of deep neural networks (DNNs) has become a key topic...
research
12/09/2021

Latent Space Explanation by Intervention

The success of deep neural nets heavily relies on their ability to encod...

Please sign up or login with your details

Forgot password? Click here to reset