All Projects → transformers-interpret → Similar Projects or Alternatives

232 Open source projects that are alternatives of or similar to transformers-interpret

Transformer-MM-Explainability
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Stars: ✭ 484 (-43.79%)
thermostat
Collection of NLP model explanations and accompanying analysis tools
Stars: ✭ 126 (-85.37%)
Mutual labels:  transformers, interpretability, captum
zennit
Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.
Stars: ✭ 57 (-93.38%)
Mutual labels:  interpretability, explainable-ai
mllp
The code of AAAI 2020 paper "Transparent Classification with Multilayer Logical Perceptrons and Random Binarization".
Stars: ✭ 15 (-98.26%)
Mutual labels:  interpretability, explainable-ai
ProtoTree
ProtoTrees: Neural Prototype Trees for Interpretable Fine-grained Image Recognition, published at CVPR2021
Stars: ✭ 47 (-94.54%)
Mutual labels:  interpretability, explainable-ai
deep-explanation-penalization
Code for using CDEP from the paper "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge" https://arxiv.org/abs/1909.13584
Stars: ✭ 110 (-87.22%)
Mutual labels:  interpretability, explainable-ai
concept-based-xai
Library implementing state-of-the-art Concept-based and Disentanglement Learning methods for Explainable AI
Stars: ✭ 41 (-95.24%)
Mutual labels:  interpretability, explainable-ai
Interpret
Fit interpretable models. Explain blackbox machine learning.
Stars: ✭ 4,352 (+405.46%)
Mutual labels:  interpretability, explainable-ai
meg
Molecular Explanation Generator
Stars: ✭ 14 (-98.37%)
Mutual labels:  interpretability, explainable-ai
ALPS 2021
XAI Tutorial for the Explainable AI track in the ALPS winter school 2021
Stars: ✭ 55 (-93.61%)
Mutual labels:  interpretability, captum
knowledge-neurons
A library for finding knowledge neurons in pretrained transformer models.
Stars: ✭ 72 (-91.64%)
Mutual labels:  transformers, interpretability
hierarchical-dnn-interpretations
Using / reproducing ACD from the paper "Hierarchical interpretations for neural network predictions" 🧠 (ICLR 2019)
Stars: ✭ 110 (-87.22%)
Mutual labels:  interpretability, explainable-ai
Basic-UI-for-GPT-J-6B-with-low-vram
A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model loading takes 12gb free ram.
Stars: ✭ 90 (-89.55%)
Mutual labels:  transformers
transformers-lightning
A collection of Models, Datasets, DataModules, Callbacks, Metrics, Losses and Loggers to better integrate pytorch-lightning with transformers.
Stars: ✭ 45 (-94.77%)
Mutual labels:  transformers
ArenaR
Data generator for Arena - interactive XAI dashboard
Stars: ✭ 28 (-96.75%)
Mutual labels:  interpretability
text
Using Transformers from HuggingFace in R
Stars: ✭ 66 (-92.33%)
Mutual labels:  transformers
RETRO-pytorch
Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch
Stars: ✭ 473 (-45.06%)
Mutual labels:  transformers
Ask2Transformers
A Framework for Textual Entailment based Zero Shot text classification
Stars: ✭ 102 (-88.15%)
Mutual labels:  transformers
Transformer-in-PyTorch
Transformer/Transformer-XL/R-Transformer examples and explanations
Stars: ✭ 21 (-97.56%)
Mutual labels:  transformers
TransQuest
Transformer based translation quality estimation
Stars: ✭ 85 (-90.13%)
Mutual labels:  transformers
Introduction-to-Deep-Learning-and-Neural-Networks-Course
Code snippets and solutions for the Introduction to Deep Learning and Neural Networks Course hosted in educative.io
Stars: ✭ 33 (-96.17%)
Mutual labels:  transformers
DataScience ArtificialIntelligence Utils
Examples of Data Science projects and Artificial Intelligence use cases
Stars: ✭ 302 (-64.92%)
Mutual labels:  explainable-ai
expmrc
ExpMRC: Explainability Evaluation for Machine Reading Comprehension
Stars: ✭ 58 (-93.26%)
Mutual labels:  explainable-ai
text-classification-transformers
Easy text classification for everyone : Bert based models via Huggingface transformers (KR / EN)
Stars: ✭ 32 (-96.28%)
Mutual labels:  transformers
mmn
Moore Machine Networks (MMN): Learning Finite-State Representations of Recurrent Policy Networks
Stars: ✭ 39 (-95.47%)
Mutual labels:  interpretability
oreilly-bert-nlp
This repository contains code for the O'Reilly Live Online Training for BERT
Stars: ✭ 19 (-97.79%)
Mutual labels:  transformers
STAM-pytorch
Implementation of STAM (Space Time Attention Model), a pure and simple attention model that reaches SOTA for video classification
Stars: ✭ 109 (-87.34%)
Mutual labels:  transformers
clip-italian
CLIP (Contrastive Language–Image Pre-training) for Italian
Stars: ✭ 113 (-86.88%)
Mutual labels:  transformers
awesome-huggingface
🤗 A list of wonderful open-source projects & applications integrated with Hugging Face libraries.
Stars: ✭ 436 (-49.36%)
Mutual labels:  transformers
adversarial-robustness-public
Code for AAAI 2018 accepted paper: "Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients"
Stars: ✭ 49 (-94.31%)
Mutual labels:  interpretability
backprop
Backprop makes it simple to use, finetune, and deploy state-of-the-art ML models.
Stars: ✭ 229 (-73.4%)
Mutual labels:  transformers
EgoCNN
Code for "Distributed, Egocentric Representations of Graphs for Detecting Critical Structures" (ICML 2019)
Stars: ✭ 16 (-98.14%)
Mutual labels:  interpretability
X-Transformer
X-Transformer: Taming Pretrained Transformers for eXtreme Multi-label Text Classification
Stars: ✭ 127 (-85.25%)
Mutual labels:  transformers
transformer generalization
The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We significantly improve the systematic generalization of transformer models on a variety of datasets using simple tricks and careful considerations.
Stars: ✭ 58 (-93.26%)
Mutual labels:  transformers
jax-models
Unofficial JAX implementations of deep learning research papers
Stars: ✭ 108 (-87.46%)
Mutual labels:  transformers
grasp
Essential NLP & ML, short & fast pure Python code
Stars: ✭ 58 (-93.26%)
Mutual labels:  explainable-ai
bert attn viz
Visualize BERT's self-attention layers on text classification tasks
Stars: ✭ 41 (-95.24%)
Mutual labels:  explainable-ai
Chinese-Minority-PLM
CINO: Pre-trained Language Models for Chinese Minority (少数民族语言预训练模型)
Stars: ✭ 133 (-84.55%)
Mutual labels:  transformers
xai-iml-sota
Interesting resources related to Explainable Artificial Intelligence, Interpretable Machine Learning, Interactive Machine Learning, Human in Loop and Visual Analytics.
Stars: ✭ 51 (-94.08%)
Mutual labels:  interpretability
ginza-transformers
Use custom tokenizers in spacy-transformers
Stars: ✭ 15 (-98.26%)
Mutual labels:  transformers
glcapsnet
Global-Local Capsule Network (GLCapsNet) is a capsule-based architecture able to provide context-based eye fixation prediction for several autonomous driving scenarios, while offering interpretability both globally and locally.
Stars: ✭ 33 (-96.17%)
Mutual labels:  interpretability
naru
Neural Relation Understanding: neural cardinality estimators for tabular data
Stars: ✭ 76 (-91.17%)
Mutual labels:  transformers
pysentimiento
A Python multilingual toolkit for Sentiment Analysis and Social NLP tasks
Stars: ✭ 274 (-68.18%)
Mutual labels:  transformers
3D-GuidedGradCAM-for-Medical-Imaging
This Repo containes the implemnetation of generating Guided-GradCAM for 3D medical Imaging using Nifti file in tensorflow 2.0. Different input files can be used in that case need to edit the input to the Guided-gradCAM model.
Stars: ✭ 60 (-93.03%)
Mutual labels:  explainable-ai
question generator
An NLP system for generating reading comprehension questions
Stars: ✭ 188 (-78.16%)
Mutual labels:  transformers
neuro-symbolic-sudoku-solver
⚙️ Solving sudoku using Deep Reinforcement learning in combination with powerful symbolic representations.
Stars: ✭ 60 (-93.03%)
Mutual labels:  explainable-ai
BottleneckTransformers
Bottleneck Transformers for Visual Recognition
Stars: ✭ 231 (-73.17%)
Mutual labels:  transformers
kernel-mod
NeurIPS 2018. Linear-time model comparison tests.
Stars: ✭ 17 (-98.03%)
Mutual labels:  interpretability
dlime experiments
In this work, we propose a deterministic version of Local Interpretable Model Agnostic Explanations (LIME) and the experimental results on three different medical datasets shows the superiority for Deterministic Local Interpretable Model-Agnostic Explanations (DLIME).
Stars: ✭ 21 (-97.56%)
Mutual labels:  explainable-ai
KB-ALBERT
KB국민은행에서 제공하는 경제/금융 도메인에 특화된 한국어 ALBERT 모델
Stars: ✭ 215 (-75.03%)
Mutual labels:  transformers
LIT
[AAAI 2022] This is the official PyTorch implementation of "Less is More: Pay Less Attention in Vision Transformers"
Stars: ✭ 79 (-90.82%)
Mutual labels:  transformers
interpretable-ml
Techniques & resources for training interpretable ML models, explaining ML models, and debugging ML models.
Stars: ✭ 17 (-98.03%)
Mutual labels:  interpretability
responsible-ai-toolbox
This project provides responsible AI user interfaces for Fairlearn, interpret-community, and Error Analysis, as well as foundational building blocks that they rely on.
Stars: ✭ 615 (-28.57%)
Mutual labels:  explainable-ai
nlp workshop odsc europe20
Extensive tutorials for the Advanced NLP Workshop in Open Data Science Conference Europe 2020. We will leverage machine learning, deep learning and deep transfer learning to learn and solve popular tasks using NLP including NER, Classification, Recommendation \ Information Retrieval, Summarization, Classification, Language Translation, Q&A and T…
Stars: ✭ 127 (-85.25%)
Mutual labels:  transformers
adaptive-wavelets
Adaptive, interpretable wavelets across domains (NeurIPS 2021)
Stars: ✭ 58 (-93.26%)
Mutual labels:  interpretability
danish transformers
A collection of Danish Transformers
Stars: ✭ 30 (-96.52%)
Mutual labels:  transformers
SnowflakeNet
(TPAMI 2022) Snowflake Point Deconvolution for Point Cloud Completion and Generation with Skip-Transformer
Stars: ✭ 74 (-91.41%)
Mutual labels:  transformers
code-transformer
Implementation of the paper "Language-agnostic representation learning of source code from structure and context".
Stars: ✭ 130 (-84.9%)
Mutual labels:  transformers
gpl
Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: Generative Pseudo Labeling for Unsupervised Domain Adaptation of Dense Retrieval" https://arxiv.org/abs/2112.07577
Stars: ✭ 216 (-74.91%)
Mutual labels:  transformers
XAIatERUM2020
Workshop: Explanation and exploration of machine learning models with R and DALEX at eRum 2020
Stars: ✭ 52 (-93.96%)
Mutual labels:  explainable-ai
1-60 of 232 similar projects