BayLIME: Bayesian Local Interpretable Model-Agnostic Explanations

by   Xingyu Zhao, et al.

A key impediment to the use of AI is the lacking of transparency, especially in safety/security critical applications. The black-box nature of AI systems prevents humans from direct explanations on how the AI makes predictions, which stimulated Explainable AI (XAI) – a research field that aims at improving the trust and transparency of AI systems. In this paper, we introduce a novel XAI technique, BayLIME, which is a Bayesian modification of the widely used XAI approach LIME. BayLIME exploits prior knowledge to improve the consistency in repeated explanations of a single prediction and also the robustness to kernel settings. Both theoretical analysis and extensive experiments are conducted to support our conclusions.


Explainable AI without Interpretable Model

Explainability has been a challenge in AI for as long as AI has existed....

TRUST XAI: Model-Agnostic Explanations for AI With a Case Study on IIoT Security

Despite AI's significant growth, its "black box" nature creates challeng...

Explainable AI by BAPC – Before and After correction Parameter Comparison

By means of a local surrogate approach, an analytical method to yield ex...

Explainable AI, but explainable to whom?

Advances in AI technologies have resulted in superior levels of AI-based...

Explainable Machine Learning for Hydrocarbon Prospect Risking

Hydrocarbon prospect risking is a critical application in geophysics pre...

Trusting the Explainers: Teacher Validation of Explainable Artificial Intelligence for Course Design

Deep learning models for learning analytics have become increasingly pop...

Utilizing Explainable AI for improving the Performance of Neural Networks

Nowadays, deep neural networks are widely used in a variety of fields th...

Please sign up or login with your details

Forgot password? Click here to reset