Hey JetsonDeep Learning based Automatic Speech Recognition with attention for the Nvidia Jetson.
Stars: ✭ 161 (-65.23%)
Punctuator2A bidirectional recurrent neural network model with attention mechanism for restoring missing punctuation in unsegmented text
Stars: ✭ 483 (+4.32%)
RnnsharpRNNSharp is a toolkit of deep recurrent neural network which is widely used for many different kinds of tasks, such as sequence labeling, sequence-to-sequence and so on. It's written by C# language and based on .NET framework 4.6 or above versions. RNNSharp supports many different types of networks, such as forward and bi-directional network, sequence-to-sequence network, and different types of layers, such as LSTM, Softmax, sampled Softmax and others.
Stars: ✭ 277 (-40.17%)
Deep learning nlpKeras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Stars: ✭ 407 (-12.1%)
datastories-semeval2017-task6Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (-95.68%)
Image Caption GeneratorA neural network to generate captions for an image using CNN and RNN with BEAM Search.
Stars: ✭ 126 (-72.79%)
Hscrf PytorchACL 2018: Hybrid semi-Markov CRF for Neural Sequence Labeling (http://aclweb.org/anthology/P18-2038)
Stars: ✭ 284 (-38.66%)
Nlp Projectsword2vec, sentence2vec, machine reading comprehension, dialog system, text classification, pretrained language model (i.e., XLNet, BERT, ELMo, GPT), sequence labeling, information retrieval, information extraction (i.e., entity, relation and event extraction), knowledge graph, text generation, network embedding
Stars: ✭ 360 (-22.25%)
Delfta Deep Learning Framework for Text
Stars: ✭ 289 (-37.58%)
Rnn For Joint NluTensorflow implementation of "Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling" (https://arxiv.org/abs/1609.01454)
Stars: ✭ 281 (-39.31%)
Neural spEnd-to-end ASR/LM implementation with PyTorch
Stars: ✭ 408 (-11.88%)
Easy TensorflowSimple and comprehensive tutorials in TensorFlow
Stars: ✭ 2,871 (+520.09%)
Sltk序列化标注工具,基于PyTorch实现BLSTM-CNN-CRF模型,CoNLL 2003 English NER测试集F1值为91.10%(word and char feature)。
Stars: ✭ 338 (-27%)
Encoder decoderFour styles of encoder decoder model by Python, Theano, Keras and Seq2Seq
Stars: ✭ 269 (-41.9%)
AttentionwalkA PyTorch Implementation of "Watch Your Step: Learning Node Embeddings via Graph Attention" (NeurIPS 2018).
Stars: ✭ 266 (-42.55%)
GansformerGenerative Adversarial Transformers
Stars: ✭ 421 (-9.07%)
Nlp TutorialsSimple implementations of NLP models. Tutorials are written in Chinese on my website https://mofanpy.com
Stars: ✭ 394 (-14.9%)
Carrot🥕 Evolutionary Neural Networks in JavaScript
Stars: ✭ 261 (-43.63%)
NagisaA Japanese tokenizer based on recurrent neural networks
Stars: ✭ 260 (-43.84%)
ReturnnThe RWTH extensible training framework for universal recurrent neural networks
Stars: ✭ 290 (-37.37%)
RmdlRMDL: Random Multimodel Deep Learning for Classification
Stars: ✭ 375 (-19.01%)
GectorOfficial implementation of the paper “GECToR – Grammatical Error Correction: Tag, Not Rewrite” // Published on BEA15 Workshop (co-located with ACL 2020) https://www.aclweb.org/anthology/2020.bea-1.16.pdf
Stars: ✭ 287 (-38.01%)
Pytorch Original TransformerMy implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (-11.23%)
QtraderReinforcement Learning for Portfolio Management
Stars: ✭ 363 (-21.6%)
Recurrent Entity NetworksTensorFlow implementation of "Tracking the World State with Recurrent Entity Networks".
Stars: ✭ 276 (-40.39%)
Mcan VqaDeep Modular Co-Attention Networks for Visual Question Answering
Stars: ✭ 273 (-41.04%)
AutonerLearning Named Entity Tagger from Domain-Specific Dictionary
Stars: ✭ 357 (-22.89%)
Abd Net[ICCV 2019] "ABD-Net: Attentive but Diverse Person Re-Identification" https://arxiv.org/abs/1908.01114
Stars: ✭ 272 (-41.25%)
Lstm Human Activity RecognitionHuman Activity Recognition example using TensorFlow on smartphone sensors dataset and an LSTM RNN. Classifying the type of movement amongst six activity categories - Guillaume Chevalier
Stars: ✭ 2,943 (+535.64%)
Ner BertBERT-NER (nert-bert) with google bert https://github.com/google-research.
Stars: ✭ 339 (-26.78%)
AbcnnImplementation of ABCNN(Attention-Based Convolutional Neural Network) on Tensorflow
Stars: ✭ 264 (-42.98%)
Ban VqaBilinear attention networks for visual question answering
Stars: ✭ 449 (-3.02%)
Transformer TensorflowTensorFlow implementation of 'Attention Is All You Need (2017. 6)'
Stars: ✭ 319 (-31.1%)
CtcwordbeamsearchConnectionist Temporal Classification (CTC) decoder with dictionary and language model for TensorFlow.
Stars: ✭ 398 (-14.04%)
Da Rnn📃 **Unofficial** PyTorch Implementation of DA-RNN (arXiv:1704.02971)
Stars: ✭ 256 (-44.71%)
ZhihuThis repo contains the source code in my personal column (https://zhuanlan.zhihu.com/zhaoyeyu), implemented using Python 3.6. Including Natural Language Processing and Computer Vision projects, such as text generation, machine translation, deep convolution GAN and other actual combat code.
Stars: ✭ 3,307 (+614.25%)
neural-financeNeural Network for HFT-trading [experimental]
Stars: ✭ 67 (-85.53%)
Neuronlp2Deep neural models for core NLP tasks (Pytorch version)
Stars: ✭ 397 (-14.25%)
DeepxiDeep Xi: A deep learning approach to a priori SNR estimation implemented in TensorFlow 2/Keras. For speech enhancement and robust ASR.
Stars: ✭ 304 (-34.34%)
RCNN MDPCode base for solving Markov Decision Processes and Reinforcement Learning problems using Recurrent Convolutional Neural Networks.
Stars: ✭ 65 (-85.96%)
ResUNetPlusPlusOfficial code for ResUNetplusplus for medical image segmentation (TensorFlow implementation) (IEEE ISM)
Stars: ✭ 69 (-85.1%)
Seq2seq SummarizerPointer-generator reinforced seq2seq summarization in PyTorch
Stars: ✭ 306 (-33.91%)
CrowdLayerA neural network layer that enables training of deep neural networks directly from crowdsourced labels (e.g. from Amazon Mechanical Turk) or, more generally, labels from multiple annotators with different biases and levels of expertise.
Stars: ✭ 45 (-90.28%)
Slot filling and intent detection of sluslot filling, intent detection, joint training, ATIS & SNIPS datasets, the Facebook’s multilingual dataset, MIT corpus, E-commerce Shopping Assistant (ECSA) dataset, CoNLL2003 NER, ELMo, BERT, XLNet
Stars: ✭ 298 (-35.64%)
sgrnnTensorflow implementation of Synthetic Gradient for RNN (LSTM)
Stars: ✭ 40 (-91.36%)
PIEFast + Non-Autoregressive Grammatical Error Correction using BERT. Code and Pre-trained models for paper "Parallel Iterative Edit Models for Local Sequence Transduction": www.aclweb.org/anthology/D19-1435.pdf (EMNLP-IJCNLP 2019)
Stars: ✭ 164 (-64.58%)