Time AttentionImplementation of RNN for Time Series prediction from the paper https://arxiv.org/abs/1704.02971
Biblosa PytorchRe-implementation of Bi-Directional Block Self-Attention for Fast and Memory-Efficient Sequence Modeling (T. Shen et al., ICLR 2018) on Pytorch.
AttentionsPyTorch implementation of some attentions for Deep Learning Researchers.
AttentionclusterTensorFlow Implementation of "Attention Clusters: Purely Attention Based Local Feature Integration for Video Classification"
Attentive Neural Processesimplementing "recurrent attentive neural processes" to forecast power usage (w. LSTM baseline, MCDropout)
DefactonlpDeFactoNLP: An Automated Fact-checking System that uses Named Entity Recognition, TF-IDF vector comparison and Decomposable Attention models.
Isab PytorchAn implementation of (Induced) Set Attention Block, from the Set Transformers paper
Nlp tensorflow projectUse tensorflow to achieve some NLP project, eg: classification chatbot ner attention QAetc.
Cell DetrOfficial and maintained implementation of the paper Attention-Based Transformers for Instance Segmentation of Cells in Microstructures [BIBM 2020].
Pytorch GatMy implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Tf Rnn AttentionTensorflow implementation of attention mechanism for text classification tasks.
VadVoice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
SimplecvreproductionReproduce simple cv project including attention module, classification, object detection, segmentation, keypoint detection, tracking 😄 etc.
Speech TransformerA PyTorch implementation of Speech Transformer, an End-to-End ASR with Transformer network on Mandarin Chinese.
Performer PytorchAn implementation of Performer, a linear attention-based transformer, in Pytorch
Punctuator2A bidirectional recurrent neural network model with attention mechanism for restoring missing punctuation in unsegmented text
Rnn NluA TensorFlow implementation of Recurrent Neural Networks for Sequence Classification and Sequence Labeling
Ban VqaBilinear attention networks for visual question answering
Mac NetworkImplementation for the paper "Compositional Attention Networks for Machine Reasoning" (Hudson and Manning, ICLR 2018)
Pytorch Original TransformerMy implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Neural spEnd-to-end ASR/LM implementation with PyTorch
Deep learning nlpKeras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Nlp TutorialsSimple implementations of NLP models. Tutorials are written in Chinese on my website https://mofanpy.com
Ner BertBERT-NER (nert-bert) with google bert https://github.com/google-research.
DeepxiDeep Xi: A deep learning approach to a priori SNR estimation implemented in TensorFlow 2/Keras. For speech enhancement and robust ASR.
Mcan VqaDeep Modular Co-Attention Networks for Visual Question Answering
Abd Net[ICCV 2019] "ABD-Net: Attentive but Diverse Person Re-Identification" https://arxiv.org/abs/1908.01114
Encoder decoderFour styles of encoder decoder model by Python, Theano, Keras and Seq2Seq
AttentionwalkA PyTorch Implementation of "Watch Your Step: Learning Node Embeddings via Graph Attention" (NeurIPS 2018).
AbcnnImplementation of ABCNN(Attention-Based Convolutional Neural Network) on Tensorflow
ResUNetPlusPlusOfficial code for ResUNetplusplus for medical image segmentation (TensorFlow implementation) (IEEE ISM)
mtad-gat-pytorchPyTorch implementation of MTAD-GAT (Multivariate Time-Series Anomaly Detection via Graph Attention Networks) by Zhao et. al (2020, https://arxiv.org/abs/2009.02040).
SBR⌛ Introducing Self-Attention to Target Attentive Graph Neural Networks (AISP '22)