mllpThe code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Stars: ✭ 15 (-99.66%)
Mutual labels: transparency, interpretability, interpretable-ai, interpretable-ml, explainable-ai, explainable-ml, xai, interpretable-machine-learning, iml, machine-learning-interpretability, explainability, interpretml interpretable-mlTechniques & resources for training interpretable ML models, explaining ML models, and debugging ML models.
Stars: ✭ 17 (-99.61%)
diabetes use caseSample use case for Xavier AI in Healthcare conference: https://www.xavierhealth.org/ai-summit-day2/
Stars: ✭ 22 (-99.49%)
zennitZennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.
Stars: ✭ 57 (-98.69%)
xai-iml-sotaInteresting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Human in Loop and Visual Analytics.
Stars: ✭ 51 (-98.83%)
ProtoTreeProtoTrees: Neural Prototype Trees for Interpretable Fine-grained Image Recognition, published at CVPR2021
Stars: ✭ 47 (-98.92%)
dlime experimentsIn this work, we propose a deterministic version of Local Interpretable Model Agnostic Explanations (LIME) and the experimental results on three different medical datasets shows the superiority for Deterministic Local Interpretable Model-Agnostic Explanations (DLIME).
Stars: ✭ 21 (-99.52%)
ArenaRData generator for Arena - interactive XAI dashboard
Stars: ✭ 28 (-99.36%)
ImodelsInterpretable ML package 🔍 for concise, transparent, and accurate predictive modeling (sklearn-compatible).
Stars: ✭ 194 (-95.54%)
concept-based-xaiLibrary implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI
Stars: ✭ 41 (-99.06%)
fastshapFast approximate Shapley values in R
Stars: ✭ 79 (-98.18%)
ml-fairness-frameworkFairPut - Machine Learning Fairness Framework with LightGBM — Explainability, Robustness, Fairness (by @firmai)
Stars: ✭ 59 (-98.64%)
adaptive-waveletsAdaptive, interpretable wavelets across domains (NeurIPS 2021)
Stars: ✭ 58 (-98.67%)
Transformer-MM-Explainability[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Stars: ✭ 484 (-88.88%)
hierarchical-dnn-interpretationsUsing / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" 🧠 (ICLR 2019)
Stars: ✭ 110 (-97.47%)
responsible-ai-toolboxThis project provides responsible AI user interfaces for Fairlearn, interpret-community, and Error Analysis, as well as foundational building blocks that they rely on.
Stars: ✭ 615 (-85.87%)
deep-explanation-penalizationCode for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" https://arxiv.org/abs/1909.13584
Stars: ✭ 110 (-97.47%)
CARLACARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithms
Stars: ✭ 166 (-96.19%)
WhiteBox-Part1In this part, I've introduced and experimented with ways to interpret and evaluate models in the field of image. (Pytorch)
Stars: ✭ 34 (-99.22%)
XaiXAI - An eXplainability toolbox for machine learning
Stars: ✭ 596 (-86.31%)
CaptumModel interpretability and understanding for PyTorch
Stars: ✭ 2,830 (-34.97%)
MindsdbPredictive AI layer for existing databases.
Stars: ✭ 4,199 (-3.52%)
ExplainxExplainable AI framework for data scientists. Explain & debug any blackbox machine learning model with a single line of code.
Stars: ✭ 196 (-95.5%)
TensorwatchDebugging, monitoring and visualization for Python Machine Learning and Data Science
Stars: ✭ 3,191 (-26.68%)
ShapA game theoretic approach to explain the output of any machine learning model.
Stars: ✭ 14,917 (+242.76%)
Hyperparameter hunterEasy hyperparameter optimization and automatic result saving across machine learning algorithms and libraries
Stars: ✭ 648 (-85.11%)
Auto ml[UNMAINTAINED] Automated machine learning for analytics & production
Stars: ✭ 1,559 (-64.18%)
thermostatCollection of NLP model explanations and accompanying analysis tools
Stars: ✭ 126 (-97.1%)
GNNLens2Visualization tool for Graph Neural Networks
Stars: ✭ 155 (-96.44%)
XAIatERUM2020Workshop: Explanation and exploration of machine learning models with R and DALEX at eRum 2020
Stars: ✭ 52 (-98.81%)
ALPS 2021XAI Tutorial for the Explainable AI track in the ALPS winter school 2021
Stars: ✭ 55 (-98.74%)
expmrcExpMRC: Explainability Evaluation for Machine Reading Comprehension
Stars: ✭ 58 (-98.67%)
mindsdb serverMindsDB server allows you to consume and expose MindsDB workflows, through http.
Stars: ✭ 3 (-99.93%)
ArtificioDeep Learning Computer Vision Algorithms for Real-World Use
Stars: ✭ 326 (-92.51%)
AutoscraperA Smart, Automatic, Fast and Lightweight Web Scraper for Python
Stars: ✭ 4,077 (-6.32%)
megMolecular Explanation Generator
Stars: ✭ 14 (-99.68%)
transformers-interpretModel explainability that works seamlessly with 🤗 transformers. Explain your transformers model in just 2 lines of code.
Stars: ✭ 861 (-80.22%)
handson-ml도서 "핸즈온 머신러닝"의 예제와 연습문제를 담은 주피터 노트북입니다.
Stars: ✭ 285 (-93.45%)
Lightnet🌓 Bringing pjreddie's DarkNet out of the shadows #yolo
Stars: ✭ 322 (-92.6%)
Intelligo🤖 Chatbot Framework for Node.js.
Stars: ✭ 347 (-92.03%)
self critical vqaCode for NeurIPS 2019 paper ``Self-Critical Reasoning for Robust Visual Question Answering''
Stars: ✭ 39 (-99.1%)
Deep XFPackage towards building Explainable Forecasting and Nowcasting Models with State-of-the-art Deep Neural Networks and Dynamic Factor Model on Time Series data sets with single line of code. Also, provides utilify facility for time-series signal similarities matching, and removing noise from timeseries signals.
Stars: ✭ 83 (-98.09%)
ShapML.jlA Julia package for interpretable machine learning with stochastic Shapley values
Stars: ✭ 63 (-98.55%)
ClaiCommand Line Artificial Intelligence or CLAI is an open-sourced project from IBM Research aimed to bring the power of AI to the command line interface.
Stars: ✭ 320 (-92.65%)
yggdrasil-decision-forestsA collection of state-of-the-art algorithms for the training, serving and interpretation of Decision Forest models.
Stars: ✭ 156 (-96.42%)
sageFor calculating global feature importance using Shapley values.
Stars: ✭ 129 (-97.04%)
shaprExplaining the output of machine learning models with more accurately estimated Shapley values
Stars: ✭ 95 (-97.82%)
global-attribution-mappingGAM (Global Attribution Mapping) explains the landscape of neural network predictions across subpopulations
Stars: ✭ 18 (-99.59%)
SHAP FOLD(Explainable AI) - Learning Non-Monotonic Logic Programs From Statistical Models Using High-Utility Itemset Mining
Stars: ✭ 35 (-99.2%)
removal-explanationsA lightweight implementation of removal-based explanations for ML models.
Stars: ✭ 46 (-98.94%)
LdaLDA topic modeling for node.js
Stars: ✭ 262 (-93.98%)
StripsAI Automated Planning with STRIPS and PDDL in Node.js
Stars: ✭ 272 (-93.75%)
PolyaxonMachine Learning Platform for Kubernetes (MLOps tools for experimentation and automation)
Stars: ✭ 2,966 (-31.85%)