Interpreting Neural Network Judgments via Minimal, Stable, and Symbolic Corrections

02/21/2018
by   Xin Zhang, et al.
0

The paper describes a new algorithm to generate minimal, stable, and symbolic corrections to an input that will cause a neural network with ReLU neurons to change its output. We argue that such a correction is a useful way to provide feedback to a user when the neural network produces an output that is different from a desired output. Our algorithm generates such a correction by solving a series of linear constraint satisfaction problems. The technique is evaluated on a neural network that has been trained to predict whether an applicant will pay a mortgage.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/25/2017

A new algorithm for constraint satisfaction problems with Maltsev templates

In this article, we provide a new algorithm for solving constraint satis...
research
02/15/2021

Scaling Up Exact Neural Network Compression by ReLU Stability

We can compress a neural network while exactly preserving its underlying...
research
06/29/2020

Scaling Symbolic Methods using Gradients for Neural Model Explanation

Symbolic techniques based on Satisfiability Modulo Theory (SMT) solvers ...
research
08/17/2019

A Symbolic Neural Network Representation and its Application to Understanding, Verifying, and Patching Network

Analysis and manipulation of trained neural networks is a challenging an...
research
07/15/2023

NeurASP: Embracing Neural Networks into Answer Set Programming

We present NeurASP, a simple extension of answer set programs by embraci...
research
07/18/2020

Abstraction based Output Range Analysis for Neural Networks

In this paper, we consider the problem of output range analysis for feed...
research
06/29/2020

A Bayesian regularization-backpropagation neural network model for peeling computations

Bayesian regularization-backpropagation neural network (BR-BPNN), a mach...

Please sign up or login with your details

Forgot password? Click here to reset