All Projects → SinaMohseni → Awesome-XAI-Evaluation

SinaMohseni / Awesome-XAI-Evaluation

Licence: other
Reference tables to introduce and organize evaluation methods and measures for explainable machine learning systems

Projects that are alternatives of or similar to Awesome-XAI-Evaluation

XAIatERUM2020
Workshop: Explanation and exploration of machine learning models with R and DALEX at eRum 2020
Stars: ✭ 52 (-8.77%)
Mutual labels:  interpretable-machine-learning
Awesome Machine Learning Interpretability
A curated list of awesome machine learning interpretability resources.
Stars: ✭ 2,404 (+4117.54%)
Mutual labels:  interpretable-machine-learning
Interpret
Fit interpretable models. Explain blackbox machine learning.
Stars: ✭ 4,352 (+7535.09%)
Mutual labels:  interpretable-machine-learning
diabetes use case
Sample use case for Xavier AI in Healthcare conference: https://www.xavierhealth.org/ai-summit-day2/
Stars: ✭ 22 (-61.4%)
Mutual labels:  interpretable-machine-learning
Layerwise-Relevance-Propagation
Implementation of Layerwise Relevance Propagation for heatmapping "deep" layers
Stars: ✭ 78 (+36.84%)
Mutual labels:  interpretable-machine-learning
ml-fairness-framework
FairPut - Machine Learning Fairness Framework with LightGBM — Explainability, Robustness, Fairness (by @firmai)
Stars: ✭ 59 (+3.51%)
Mutual labels:  interpretable-machine-learning
fastshap
Fast approximate Shapley values in R
Stars: ✭ 79 (+38.6%)
Mutual labels:  interpretable-machine-learning
ProtoTree
ProtoTrees: Neural Prototype Trees for Interpretable Fine-grained Image Recognition, published at CVPR2021
Stars: ✭ 47 (-17.54%)
Mutual labels:  interpretable-machine-learning
ShapleyExplanationNetworks
Implementation of the paper "Shapley Explanation Networks"
Stars: ✭ 62 (+8.77%)
Mutual labels:  interpretable-machine-learning
ShapML.jl
A Julia package for interpretable machine learning with stochastic Shapley values
Stars: ✭ 63 (+10.53%)
Mutual labels:  interpretable-machine-learning
ConceptBottleneck
Concept Bottleneck Models, ICML 2020
Stars: ✭ 91 (+59.65%)
Mutual labels:  interpretable-machine-learning
interpretable-ml
Techniques & resources for training interpretable ML models, explaining ML models, and debugging ML models.
Stars: ✭ 17 (-70.18%)
Mutual labels:  interpretable-machine-learning
mllp
The code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Stars: ✭ 15 (-73.68%)
Mutual labels:  interpretable-machine-learning
xai-iml-sota
Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Human in Loop and Visual Analytics.
Stars: ✭ 51 (-10.53%)
Mutual labels:  interpretable-machine-learning

Awesome-XAI-Evaluation Awesome

An awesome and organized multidisciplinary reference list of evaluation measures and methods for explainable machine learning (XAI) algorithms and systems. If you need more details and descriptions, you can read the full paper or visit my page for more resources!

How to Evaluate XAI?

We reviewed XAI-related research to organize different XAI design goals and evaluation measures. This awesome-list presents our categorization of selected existing design and evaluation methods that organizes literature along with three perspectives: design goals, evaluation methods, and targeted users of the XAI system. We provide summarized, ready-to-use tables of evaluation methods and recommendations for different goals in XAI research.

users 1

Citation

Description and details in this paper: https://arxiv.org/pdf/1811.11839.pdf

@article{mohseni2018multidisciplinary,
  title={A Multidisciplinary Survey and Framework for Design and Evaluation of Explainable AI Systems},
  author={Mohseni, Sina and Zarei, Niloofar and Ragan, Eric D},
  journal={arXiv preprint arXiv:1811.11839},
  year={2018}
}

Evaluation Measures

  • Computational Measures
    • M1: Fidelity of Interpretability Method
    • M2: Model Trustworthiness
  • Human-grounded Measures
    • M3: Human-machine Task Performance
    • M4: User Mental Model
    • M5: User Trust and Reliance
    • M6: Explanation Usefulness and Satisfaction

M1: Fidelity of Interpretability Technique

Paper Evaluation Method
The Building Blocks of Interpretability Sanity Check Experiment
Graying the black box: Understanding DQNs Sanity Check Experiment
Visualizing deep neural network decisions: Prediction difference analysis Sanity Check Experiment
Understanding neural networks through deep visualization Sanity Check Experiment
The (Un)reliability of saliency methods Sanity Check Experiment
Improving the adversarial robustness and interpretability of deep neural networks by regularizing their input gradients Sanity Check Experiment
Why Should I Trust You? Explaining the Predictions of Any Classifier Simulated Experiment
Anchors: High-precision model-agnostic explanations Simulated Experiment
Evaluating the visualization of what a deep neural network has learned Comparative Evaluation
Right for the Right Reasons: Training Differentiable Models by Constraining their Explanations Comparative Evaluation
On the Robustness of Interpretability Methods Sanity Check Experiment
Human-grounded Evaluations of Explanation Methods for Text Classification Human-grounded Evaluation
Sanity Checks for Saliency Metrics Sanity Check Experiment
Benchmarking Attribution Methods with Relative Feature Importance Comparative Evaluation
Many Faces of Feature Importance: Comparing Built-in and Post-hoc Feature Importance in Text Classification Comparative Evaluation
Towards Ground Truth Evaluation of Visual Explanations Feature-based Baseline
Evaluating Recurrent Neural Network Explanations Sanity Check
Evaluating neural network explanation methods using hybrid documents and morphosyntactic agreement Comparative Evaluation

M2: Model Trustworthiness

Paper Evaluation Method
A Human-Grounded Evaluation Benchmark for Local Explanations of Machine Learning Human-grounded Baseline
A unified approach to interpreting model predictions Human Judgment
Quantifying Interpretability and Trust in Machine Learning Systems Human Judgment
Human attention in visual question answering: Do humans and deep networks look at the same regions? Human-grounded Baseline
Visualizing and understanding convolutional networks Debugging model and training
Towards Explanation of DNN-based Prediction with Guided Feature Inversion Human-grounded Baseline
Explainable Deep Classification Models for Domain Generalization Human Judgment
Score-CAM: Improved Visual Explanations Via Score-Weighted Class Activation Mapping Human-grounded Baseline
Human-in-the-Loop Interpretability Prior Human Judgment

M3: Human-machine Task Performance

Paper Evaluation Method
Explanatory debugging: Supporting end-user debugging of machine-learned programs Task Performance, Task Throughput
Why and why not explanations improve the intelligibility of context-aware intelligent systems Task Performance, Task Throughput
ActiVis: Visual Exploration of Industry-Scale Deep Neural Network Models Task Performance
You are the only possible oracle: Effective test selection for end users of interactive machine learning systems Task Performance, Model Failure Prediction
Interpretable decision sets: A joint framework for description and prediction Task Throughput
A Workflow for Visual Diagnostics of Binary Classifiers using Instance-Level Explanations Model Failure Prediction
Interacting meaningfully with machine learning systems: Three experiments Model Failure Prediction, Model Accuracy
Why should I you?: Explaining the predictions of any classifier Model Accuracy
Principles of explanatory debugging to personalize interactive machine learning Model Accuracy
Towards better analysis of deep convolutional neural networks Model Accuracy
Deepeyes: Progressive visual analytics for designing deep neural networks Model Accuracy
Topicpanorama: A full picture of relevant topics Model Tuning and Selection
Updates in human-ai teams: Understanding and addressing the performance/compatibility tradeoff Success Rate
Human Evaluation of Models Built for Interpretability User Prediction Accuracy
Evaluating Visual Explanations for Similarity-Based Recommendations: User Perception and Performance --
Can You Explain That? Lucid Explanations Help Human-AI Collaborative Image Retrieval --
Leveraging Rationales to Improve Human Task Performance --

M4: User Mental Model

Paper Evaluation Method
Updates in human-ai teams: Understanding and addressing the performance/compatibility tradeoff Success Rate
Intellingo: An Intelligible Translation Environment --
Human Evaluation of Models Built for Interpretability User Prediction Accuracy
Evaluating Visual Explanations for Similarity-Based Recommendations: User Perception and Performance --
When People and Algorithms Meet: User-reported Problems in Intelligent Everyday Applications --
Towards Accountable AI: Hybrid Human-Machine Analyses for Characterizing System Failure --
Beyond Accuracy: The Role of Mental Models in Human-AI Team Performance User Success Rate
Can You Explain That? Lucid Explanations Help Human-AI Collaborative Image Retrieval --
A Case for Backward Compatibility for Human-AI Teams --
Leveraging Rationales to Improve Human Task Performance --

M5: User Trust and Reliance

Paper Evaluation Method
The Impact of Placebic Explanations on Trust in Intelligent Systems Agreement Rate
The Effects of Meaningful and Meaningless Explanations on Trust and Perceived System Accuracy in Intelligent Systems User Perceived Accuracy
I Drive — You Trust: Explaining Driving Behavior Of Autonomous Cars Subjective Rating
The role of explanations on trust and reliance in clinical decision support systems Agreement Rate
Understanding the Effect of Accuracy on Trust in Machine Learning Models Agreement and Switch Rate
How much information?: Effects of transparency on trust in an algorithmic interface Subjective Rating
“How do I fool you?”: Manipulating User Trust via Misleading Black Box Explanations --
Do I Trust My Machine Teammate? An Investigation from Perception to Decision --
Evaluating Effects of User Experience and System Transparency on Trust in Automation --
Trust Calibration within a Human-Robot Team: Comparing Automatically Generated Explanations --
User trust in intelligent systems: A journey over time Subjective Rating
What Does Explainable AI Really Mean? A New Conceptualization of Perspectives --
The effects of example-based explanations in a machine learning interface --
Let Me Explain: Impact of Personal and Impersonal Explanations on Trust in Recommender Systems --
Trust Dynamics in Human Autonomous Vehicle Interaction: A Review of Trust Models --
Effects of Model Confidence and Explanation on Accuracy and Trust Calibration Agreement and Switch Rate
“How do I fool you?”: Manipulating User Trust via Misleading Black Box Explanations Subjective Rating

M6: Explanation Usefulness and Satisfaction

Paper Evaluation Method
Are explanations always important? a study of deployed, low-cost intelligent interactive systems Interview and Self-report
Assessing demand for intelligibility in context-aware applications Interview and Self-report
How should I explain? A comparison of different explanation types for recommender systems Interview, Self-report, User Learning duration
Why and why not explanations improve the intelligibility of context-aware intelligent systems Interview and Self-report
Intellingo: An Intelligible Translation Environment Likert-scale Questionnaire
Human Evaluation of Models Built for Interpretability Likert-scale Questionnaire
Intellingo: An Intelligible Translation Environment Engagement with Explanations
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].