InterpretFit interpretable models. Explain blackbox machine learning.
Stars: ✭ 4,352 (+15442.86%)
mllpThe code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Stars: ✭ 15 (-46.43%)
zennitZennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.
Stars: ✭ 57 (+103.57%)
adaptive-waveletsAdaptive, interpretable wavelets across domains (NeurIPS 2021)
Stars: ✭ 58 (+107.14%)
concept-based-xaiLibrary implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI
Stars: ✭ 41 (+46.43%)
diabetes use caseSample use case for Xavier AI in Healthcare conference: https://www.xavierhealth.org/ai-summit-day2/
Stars: ✭ 22 (-21.43%)
xai-iml-sotaInteresting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Human in Loop and Visual Analytics.
Stars: ✭ 51 (+82.14%)
interpretable-mlTechniques & resources for training interpretable ML models, explaining ML models, and debugging ML models.
Stars: ✭ 17 (-39.29%)
ShapA game theoretic approach to explain the output of any machine learning model.
Stars: ✭ 14,917 (+53175%)
ProtoTreeProtoTrees: Neural Prototype Trees for Interpretable Fine-grained Image Recognition, published at CVPR2021
Stars: ✭ 47 (+67.86%)
thermostatCollection of NLP model explanations and accompanying analysis tools
Stars: ✭ 126 (+350%)
deep-explanation-penalizationCode for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" https://arxiv.org/abs/1909.13584
Stars: ✭ 110 (+292.86%)
Transformer-MM-Explainability[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Stars: ✭ 484 (+1628.57%)
ALPS 2021XAI Tutorial for the Explainable AI track in the ALPS winter school 2021
Stars: ✭ 55 (+96.43%)
sageFor calculating global feature importance using Shapley values.
Stars: ✭ 129 (+360.71%)
hierarchical-dnn-interpretationsUsing / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" 🧠 (ICLR 2019)
Stars: ✭ 110 (+292.86%)
removal-explanationsA lightweight implementation of removal-based explanations for ML models.
Stars: ✭ 46 (+64.29%)
GNNLens2Visualization tool for Graph Neural Networks
Stars: ✭ 155 (+453.57%)
BreakdownModel Agnostics breakDown plots
Stars: ✭ 93 (+232.14%)
DeepAccNetPytorch/Python3 implementation of DeepAccNet, protein model accuracy evaluator.
Stars: ✭ 57 (+103.57%)
Interpretability By PartsCode repository for "Interpretable and Accurate Fine-grained Recognition via Region Grouping", CVPR 2020 (Oral)
Stars: ✭ 88 (+214.29%)
spatio-temporal-brainA Deep Graph Neural Network Architecture for Modelling Spatio-temporal Dynamics in rs-fMRI Data
Stars: ✭ 22 (-21.43%)
Cnn Interpretability🏥 Visualizing Convolutional Networks for MRI-based Diagnosis of Alzheimer’s Disease
Stars: ✭ 68 (+142.86%)
Pycebox⬛ Python Individual Conditional Expectation Plot Toolbox
Stars: ✭ 101 (+260.71%)
finmathThe collections of simple, weighted, exponential, smoothed moving averages.
Stars: ✭ 49 (+75%)
kernel-modNeurIPS 2018. Linear-time model comparison tests.
Stars: ✭ 17 (-39.29%)
CxplainCausal Explanation (CXPlain) is a method for explaining the predictions of any machine-learning model.
Stars: ✭ 84 (+200%)
Torch CamClass activation maps for your PyTorch models (CAM, Grad-CAM, Grad-CAM++, Smooth Grad-CAM++, Score-CAM, SS-CAM, IS-CAM, XGrad-CAM)
Stars: ✭ 249 (+789.29%)
CaptumModel interpretability and understanding for PyTorch
Stars: ✭ 2,830 (+10007.14%)
AthenaAutomatic equation building and curve fitting. Runs on Tensorflow. Built for academia and research.
Stars: ✭ 57 (+103.57%)
Text nnText classification models. Used a submodule for other projects.
Stars: ✭ 55 (+96.43%)
ExplainxExplainable AI framework for data scientists. Explain & debug any blackbox machine learning model with a single line of code.
Stars: ✭ 196 (+600%)
TrelawneyGeneral Interpretability Package
Stars: ✭ 55 (+96.43%)
ContrastiveexplanationContrastive Explanation (Foil Trees), developed at TNO/Utrecht University
Stars: ✭ 36 (+28.57%)
Symbolic MetamodelingCodebase for "Demystifying Black-box Models with Symbolic Metamodels", NeurIPS 2019.
Stars: ✭ 29 (+3.57%)
AlibiAlgorithms for monitoring and explaining machine learning models
Stars: ✭ 924 (+3200%)
expmrcExpMRC: Explainability Evaluation for Machine Reading Comprehension
Stars: ✭ 58 (+107.14%)
XAIatERUM2020Workshop: Explanation and exploration of machine learning models with R and DALEX at eRum 2020
Stars: ✭ 52 (+85.71%)
ImodelsInterpretable ML package 🔍 for concise, transparent, and accurate predictive modeling (sklearn-compatible).
Stars: ✭ 194 (+592.86%)
Grad Cam[ICCV 2017] Torch code for Grad-CAM
Stars: ✭ 891 (+3082.14%)
DalexmoDel Agnostic Language for Exploration and eXplanation
Stars: ✭ 795 (+2739.29%)
Pyss3A Python package implementing a new machine learning model for text classification with visualization tools for Explainable AI
Stars: ✭ 191 (+582.14%)
Tf ExplainInterpretability Methods for tf.keras models with Tensorflow 2.x
Stars: ✭ 780 (+2685.71%)
mindsdb nativeMachine Learning in one line of code
Stars: ✭ 34 (+21.43%)
Ad examplesA collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
Stars: ✭ 641 (+2189.29%)
XaiXAI - An eXplainability toolbox for machine learning
Stars: ✭ 596 (+2028.57%)
FlashtorchVisualization toolkit for neural networks in PyTorch! Demo -->
Stars: ✭ 561 (+1903.57%)
Xai resourcesInteresting resources related to XAI (Explainable Artificial Intelligence)
Stars: ✭ 553 (+1875%)
mindsdb serverMindsDB server allows you to consume and expose MindsDB workflows, through http.
Stars: ✭ 3 (-89.29%)
Modelstudio📍 Interactive Studio for Explanatory Model Analysis
Stars: ✭ 163 (+482.14%)
Interpretable machine learning with pythonExamples of techniques for training interpretable ML models, explaining ML models, and debugging ML models for accuracy, discrimination, and security.
Stars: ✭ 530 (+1792.86%)