KashgariKashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (+5097.67%)
Global Self Attention NetworkA Pytorch implementation of Global Self-Attention Network, a fully-attention backbone for vision tasks
Stars: ✭ 64 (+48.84%)
2D-LSTM-Seq2SeqPyTorch implementation of a 2D-LSTM Seq2Seq Model for NMT.
Stars: ✭ 25 (-41.86%)
Yolov4 PytorchThis is a pytorch repository of YOLOv4, attentive YOLOv4 and mobilenet YOLOv4 with PASCAL VOC and COCO
Stars: ✭ 1,070 (+2388.37%)
Deep News SummarizationNews summarization using sequence to sequence model with attention in TensorFlow.
Stars: ✭ 167 (+288.37%)
AttnSleep[IEEE TNSRE] "An Attention-based Deep Learning Approach for Sleep Stage Classification with Single-Channel EEG"
Stars: ✭ 76 (+76.74%)
Im2LaTeXAn implementation of the Show, Attend and Tell paper in Tensorflow, for the OpenAI Im2LaTeX suggested problem
Stars: ✭ 16 (-62.79%)
TaLKConvolutionsOfficial PyTorch implementation of Time-aware Large Kernel (TaLK) Convolutions (ICML 2020)
Stars: ✭ 26 (-39.53%)
Nuts自然语言处理常见任务(主要包括文本分类,序列标注,自动问答等)解决方案试验田
Stars: ✭ 21 (-51.16%)
staginSTAGIN: Spatio-Temporal Attention Graph Isomorphism Network
Stars: ✭ 34 (-20.93%)
Attentive Neural Processesimplementing "recurrent attentive neural processes" to forecast power usage (w. LSTM baseline, MCDropout)
Stars: ✭ 33 (-23.26%)
Neural-Machine-TranslationSeveral basic neural machine translation models implemented by PyTorch & TensorFlow
Stars: ✭ 29 (-32.56%)
Pytorch GatMy implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
Stars: ✭ 908 (+2011.63%)
datastories-semeval2017-task6Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (-53.49%)
Tf Rnn AttentionTensorflow implementation of attention mechanism for text classification tasks.
Stars: ✭ 735 (+1609.3%)
Awesome ChatbotAwesome Chatbot Projects,Corpus,Papers,Tutorials.Chinese Chatbot =>:
Stars: ✭ 1,785 (+4051.16%)
VadVoice activity detection (VAD) toolkit including DNN, bDNN, LSTM and ACAM based VAD. We also provide our directly recorded dataset.
Stars: ✭ 622 (+1346.51%)
deep-transTransliterating English to Hindi using Recurrent Neural Networks
Stars: ✭ 44 (+2.33%)
torch-asgAuto Segmentation Criterion (ASG) implemented in pytorch
Stars: ✭ 42 (-2.33%)
Performer PytorchAn implementation of Performer, a linear attention-based transformer, in Pytorch
Stars: ✭ 546 (+1169.77%)
attention-ocrA pytorch implementation of the attention based ocr
Stars: ✭ 44 (+2.33%)
ai-visual-storytelling-seq2seqImplementation of seq2seq model for Visual Storytelling Challenge (VIST) http://visionandlanguage.net/VIST/index.html
Stars: ✭ 50 (+16.28%)
Rnn NluA TensorFlow implementation of Recurrent Neural Networks for Sequence Classification and Sequence Labeling
Stars: ✭ 463 (+976.74%)
DeepcorrectText and Punctuation correction with Deep Learning
Stars: ✭ 105 (+144.19%)
NiuTrans.NMTA Fast Neural Machine Translation System. It is developed in C++ and resorts to NiuTensor for fast tensor APIs.
Stars: ✭ 112 (+160.47%)
GansformerGenerative Adversarial Transformers
Stars: ✭ 421 (+879.07%)
Tacotron PytorchA Pytorch Implementation of Tacotron: End-to-end Text-to-speech Deep-Learning Model
Stars: ✭ 104 (+141.86%)
chatbot🤖️ 基于 PyTorch 的任务型聊天机器人(支持私有部署和 docker 部署的 Chatbot)
Stars: ✭ 77 (+79.07%)
sentencepiece-jniJava JNI wrapper for SentencePiece: unsupervised text tokenizer for Neural Network-based text generation.
Stars: ✭ 26 (-39.53%)
TextSummareimplementing Neural Summarization by Extracting Sentences and Words
Stars: ✭ 16 (-62.79%)
neural-chatAn AI chatbot using seq2seq
Stars: ✭ 30 (-30.23%)
CakechatCakeChat: Emotional Generative Dialog System
Stars: ✭ 1,361 (+3065.12%)
visualizationa collection of visualization function
Stars: ✭ 189 (+339.53%)
Im2latexPytorch implemention of Deep CNN Encoder + LSTM Decoder with Attention for Image to Latex
Stars: ✭ 84 (+95.35%)
Abd Net[ICCV 2019] "ABD-Net: Attentive but Diverse Person Re-Identification" https://arxiv.org/abs/1908.01114
Stars: ✭ 272 (+532.56%)
convolutional seq2seqfairseq: Convolutional Sequence to Sequence Learning (Gehring et al. 2017) by Chainer
Stars: ✭ 63 (+46.51%)
AbcnnImplementation of ABCNN(Attention-Based Convolutional Neural Network) on Tensorflow
Stars: ✭ 264 (+513.95%)
zeroZero -- A neural machine translation system
Stars: ✭ 121 (+181.4%)
how attentive are gatsCode for the paper "How Attentive are Graph Attention Networks?" (ICLR'2022)
Stars: ✭ 200 (+365.12%)
DLCV2018SPRINGDeep Learning for Computer Vision (CommE 5052) in NTU
Stars: ✭ 38 (-11.63%)
dtsA Keras library for multi-step time-series forecasting.
Stars: ✭ 130 (+202.33%)
LNSwipeCell一套友好的、方便集成的针对cell的左滑编辑功能!
Stars: ✭ 16 (-62.79%)
NLP-paper🎨 🎨NLP 自然语言处理教程 🎨🎨 https://dataxujing.github.io/NLP-paper/
Stars: ✭ 23 (-46.51%)
dreyeve[TPAMI 2018] Predicting the Driver’s Focus of Attention: the DR(eye)VE Project. A deep neural network learnt to reproduce the human driver focus of attention (FoA) in a variety of real-world driving scenarios.
Stars: ✭ 88 (+104.65%)
MoChA-pytorchPyTorch Implementation of "Monotonic Chunkwise Attention" (ICLR 2018)
Stars: ✭ 65 (+51.16%)
ttslearnttslearn: Library for Pythonで学ぶ音声合成 (Text-to-speech with Python)
Stars: ✭ 158 (+267.44%)
Seq2Seq-chatbotTensorFlow Implementation of Twitter Chatbot
Stars: ✭ 18 (-58.14%)
GAN-RNN Timeseries-imputationRecurrent GAN for imputation of time series data. Implemented in TensorFlow 2 on Wikipedia Web Traffic Forecast dataset from Kaggle.
Stars: ✭ 107 (+148.84%)