All Projects → minimal-nmt → Similar Projects or Alternatives

455 Open source projects that are alternatives of or similar to minimal-nmt

seq3
Source code for the NAACL 2019 paper "SEQ^3: Differentiable Sequence-to-Sequence-to-Sequence Autoencoder for Unsupervised Abstractive Sentence Compression"
Stars: ✭ 121 (+236.11%)
Mutual labels:  seq2seq
Optic-Disc-Unet
Attention Unet model with post process for retina optic disc segmention
Stars: ✭ 77 (+113.89%)
Mutual labels:  attention-mechanism
Machine-Translation-Hindi-to-english-
Machine translation is the task of converting one language to other. Unlike the traditional phrase-based translation system which consists of many small sub-components that are tuned separately, neural machine translation attempts to build and train a single, large neural network that reads a sentence and outputs a correct translation.
Stars: ✭ 19 (-47.22%)
Mutual labels:  attention-mechanism
Transformer Temporal Tagger
Code and data form the paper BERT Got a Date: Introducing Transformers to Temporal Tagging
Stars: ✭ 55 (+52.78%)
Mutual labels:  seq2seq
CVAE Dial
CVAE_XGate model in paper "Xu, Dusek, Konstas, Rieser. Better Conversations by Modeling, Filtering, and Optimizing for Coherence and Diversity"
Stars: ✭ 16 (-55.56%)
Mutual labels:  seq2seq
Hierarchical-Word-Sense-Disambiguation-using-WordNet-Senses
Word Sense Disambiguation using Word Specific models, All word models and Hierarchical models in Tensorflow
Stars: ✭ 33 (-8.33%)
Mutual labels:  attention-mechanism
lang2logic-PyTorch
PyTorch port of the paper "Language to Logical Form with Neural Attention"
Stars: ✭ 34 (-5.56%)
Mutual labels:  seq2seq
Seq2Seq-chatbot
TensorFlow Implementation of Twitter Chatbot
Stars: ✭ 18 (-50%)
Mutual labels:  seq2seq
tensorflow-ml-nlp-tf2
텐서플로2와 머신러닝으로 시작하는 자연어처리 (로지스틱회귀부터 BERT와 GPT3까지) 실습자료
Stars: ✭ 245 (+580.56%)
Mutual labels:  seq2seq
bytenet translation
A TensorFlow Implementation of Machine Translation In Neural Machine Translation in Linear Time
Stars: ✭ 60 (+66.67%)
visdial
Visual Dialog: Light-weight Transformer for Many Inputs (ECCV 2020)
Stars: ✭ 27 (-25%)
Mutual labels:  attention-mechanism
seq2seq-autoencoder
Theano implementation of Sequence-to-Sequence Autoencoder
Stars: ✭ 12 (-66.67%)
Mutual labels:  seq2seq
STAM-pytorch
Implementation of STAM (Space Time Attention Model), a pure and simple attention model that reaches SOTA for video classification
Stars: ✭ 109 (+202.78%)
Mutual labels:  attention-mechanism
NiuTrans.NMT
A Fast Neural Machine Translation System. It is developed in C++ and resorts to NiuTensor for fast tensor APIs.
Stars: ✭ 112 (+211.11%)
Video-Description-with-Spatial-Temporal-Attention
[ACM MM 2017 & IEEE TMM 2020] This is the Theano code for the paper "Video Description with Spatial Temporal Attention"
Stars: ✭ 53 (+47.22%)
Mutual labels:  attention-mechanism
tensorflow-chatbot-chinese
網頁聊天機器人 | tensorflow implementation of seq2seq model with bahdanau attention and Word2Vec pretrained embedding
Stars: ✭ 50 (+38.89%)
Mutual labels:  seq2seq
Attention mechanism-event-extraction
Attention mechanism in CNNs to extract events of interest
Stars: ✭ 17 (-52.78%)
Mutual labels:  attention-mechanism
chinese ancient poetry
seq2seq attention tensorflow textrank context
Stars: ✭ 30 (-16.67%)
Mutual labels:  seq2seq
SiGAT
source code for signed graph attention networks (ICANN2019) & SDGNN (AAAI2021)
Stars: ✭ 37 (+2.78%)
Mutual labels:  attention-mechanism
TianChi AIEarth
TianChi AIEarth Contest Solution
Stars: ✭ 57 (+58.33%)
Mutual labels:  attention-mechanism
Multi-task-Conditional-Attention-Networks
A prototype version of our submitted paper: Conversion Prediction Using Multi-task Conditional Attention Networks to Support the Creation of Effective Ad Creatives.
Stars: ✭ 21 (-41.67%)
Mutual labels:  attention-mechanism
RETRO-pytorch
Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch
Stars: ✭ 473 (+1213.89%)
Mutual labels:  attention-mechanism
question-generation
Neural Models for Key Phrase Detection and Question Generation
Stars: ✭ 29 (-19.44%)
Mutual labels:  attention-mechanism
Transformers-RL
An easy PyTorch implementation of "Stabilizing Transformers for Reinforcement Learning"
Stars: ✭ 107 (+197.22%)
Mutual labels:  attention-mechanism
vat nmt
Implementation of "Effective Adversarial Regularization for Neural Machine Translation", ACL 2019
Stars: ✭ 22 (-38.89%)
chatbot
一个基于深度学习的中文聊天机器人,这里有详细的教程与代码,每份代码都有详细的注释,作为学习是美好的选择。A Chinese chatbot based on deep learning.
Stars: ✭ 94 (+161.11%)
Mutual labels:  seq2seq
axial-attention
Implementation of Axial attention - attending to multi-dimensional data efficiently
Stars: ✭ 245 (+580.56%)
Mutual labels:  attention-mechanism
DARNN
A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction
Stars: ✭ 90 (+150%)
Mutual labels:  attention-mechanism
lstm-attention
Attention-based bidirectional LSTM for Classification Task (ICASSP)
Stars: ✭ 87 (+141.67%)
Mutual labels:  attention-mechanism
Neural-Machine-Translation
Several basic neural machine translation models implemented by PyTorch & TensorFlow
Stars: ✭ 29 (-19.44%)
Naver-AI-Hackathon-Speech
2019 Clova AI Hackathon : Speech - Rank 12 / Team Kai.Lib
Stars: ✭ 26 (-27.78%)
Mutual labels:  seq2seq
SRB
Code for "Improving Semantic Relevance for Sequence-to-Sequence Learning of Chinese Social Media Text Summarization"
Stars: ✭ 41 (+13.89%)
Mutual labels:  seq2seq
datastories-semeval2017-task6
Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (-44.44%)
Mutual labels:  attention-mechanism
sentence2vec
Deep sentence embedding using Sequence to Sequence learning
Stars: ✭ 23 (-36.11%)
Mutual labels:  seq2seq
stanford-cs231n-assignments-2020
This repository contains my solutions to the assignments for Stanford's CS231n "Convolutional Neural Networks for Visual Recognition" (Spring 2020).
Stars: ✭ 84 (+133.33%)
Mutual labels:  attention-mechanism
sentencepiece-jni
Java JNI wrapper for SentencePiece: unsupervised text tokenizer for Neural Network-based text generation.
Stars: ✭ 26 (-27.78%)
skt
Sanskrit compound segmentation using seq2seq model
Stars: ✭ 21 (-41.67%)
Mutual labels:  seq2seq
chatbot
Seq2Seq Chatbot with attention mechanism
Stars: ✭ 19 (-47.22%)
Mutual labels:  seq2seq
rnn-text-classification-tf
Tensorflow implementation of Attention-based Bidirectional RNN text classification.
Stars: ✭ 26 (-27.78%)
Mutual labels:  attention-mechanism
Speech recognition with tensorflow
Implementation of a seq2seq model for Speech Recognition using the latest version of TensorFlow. Architecture similar to Listen, Attend and Spell.
Stars: ✭ 253 (+602.78%)
Mutual labels:  seq2seq
Tensorflow novelist
模仿莎士比亚创作戏剧!屌炸天的是还能创作金庸武侠小说!快star,保持更新!!
Stars: ✭ 244 (+577.78%)
Mutual labels:  seq2seq
Pytorch Batch Attention Seq2seq
PyTorch implementation of batched bi-RNN encoder and attention-decoder.
Stars: ✭ 245 (+580.56%)
Mutual labels:  seq2seq
Mead Baseline
Deep-Learning Model Exploration and Development for NLP
Stars: ✭ 238 (+561.11%)
Mutual labels:  seq2seq
long-short-transformer
Implementation of Long-Short Transformer, combining local and global inductive biases for attention over long sequences, in Pytorch
Stars: ✭ 103 (+186.11%)
Mutual labels:  attention-mechanism
CrabNet
Predict materials properties using only the composition information!
Stars: ✭ 57 (+58.33%)
Mutual labels:  attention-mechanism
kospeech
Open-Source Toolkit for End-to-End Korean Automatic Speech Recognition leveraging PyTorch and Hydra.
Stars: ✭ 456 (+1166.67%)
Mutual labels:  seq2seq
Debug seq2seq
[unmaintained] Make seq2seq for keras work
Stars: ✭ 233 (+547.22%)
Mutual labels:  seq2seq
Text summarization with tensorflow
Implementation of a seq2seq model for summarization of textual data. Demonstrated on amazon reviews, github issues and news articles.
Stars: ✭ 226 (+527.78%)
Mutual labels:  seq2seq
LanguageModel-using-Attention
Pytorch implementation of a basic language model using Attention in LSTM network
Stars: ✭ 27 (-25%)
Mutual labels:  attention-mechanism
Nlp Tools
😋本项目旨在通过Tensorflow基于BiLSTM+CRF实现中文分词、词性标注、命名实体识别(NER)。
Stars: ✭ 225 (+525%)
Mutual labels:  seq2seq
Paddlenlp
NLP Core Library and Model Zoo based on PaddlePaddle 2.0
Stars: ✭ 212 (+488.89%)
Mutual labels:  seq2seq
MT-Preparation
Machine Translation (MT) Preparation Scripts
Stars: ✭ 15 (-58.33%)
En-transformer
Implementation of E(n)-Transformer, which extends the ideas of Welling's E(n)-Equivariant Graph Neural Network to attention
Stars: ✭ 131 (+263.89%)
Mutual labels:  attention-mechanism
Natural Language Processing With Tensorflow
Natural Language Processing with TensorFlow, published by Packt
Stars: ✭ 222 (+516.67%)
Mutual labels:  seq2seq
Headliner
🏖 Easy training and deployment of seq2seq models.
Stars: ✭ 221 (+513.89%)
Mutual labels:  seq2seq
convolutional seq2seq
fairseq: Convolutional Sequence to Sequence Learning (Gehring et al. 2017) by Chainer
Stars: ✭ 63 (+75%)
Mutual labels:  seq2seq
Screenshot To Code
A neural network that transforms a design mock-up into a static website.
Stars: ✭ 13,561 (+37569.44%)
Mutual labels:  seq2seq
zero
Zero -- A neural machine translation system
Stars: ✭ 121 (+236.11%)
Adversarial-Learning-for-Generative-Conversational-Agents
This repository contains a new adversarial training method for Generative Conversational Agents
Stars: ✭ 71 (+97.22%)
Mutual labels:  seq2seq
Compact-Global-Descriptor
Pytorch implementation of "Compact Global Descriptor for Neural Networks" (CGD).
Stars: ✭ 22 (-38.89%)
Mutual labels:  attention-mechanism
61-120 of 455 similar projects