All Projects → elbayadm → Attn2d

elbayadm / Attn2d

Licence: mit
Pervasive Attention: 2D Convolutional Networks for Sequence-to-Sequence Prediction

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Attn2d

Transformer Clinic
Understanding the Difficulty of Training Transformers
Stars: ✭ 179 (-62.32%)
Mutual labels:  nmt
Quality-Estimation2
机器翻译子任务-翻译质量评价-在BERT模型后面加上Bi-LSTM进行fine-tuning
Stars: ✭ 31 (-93.47%)
Mutual labels:  nmt
Tf Seq2seq
Sequence to sequence learning using TensorFlow.
Stars: ✭ 387 (-18.53%)
Mutual labels:  nmt
nmt
Network mapping tool
Stars: ✭ 16 (-96.63%)
Mutual labels:  nmt
extractive rc by runtime mt
Code and datasets of "Multilingual Extractive Reading Comprehension by Runtime Machine Translation"
Stars: ✭ 36 (-92.42%)
Mutual labels:  nmt
pynmt
a simple and complete pytorch implementation of neural machine translation system
Stars: ✭ 13 (-97.26%)
Mutual labels:  nmt
Subword Nmt
Unsupervised Word Segmentation for Neural Machine Translation and Text Generation
Stars: ✭ 1,819 (+282.95%)
Mutual labels:  nmt
Deep learning nlp
Keras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Stars: ✭ 407 (-14.32%)
Mutual labels:  nmt
RNNSearch
An implementation of attention-based neural machine translation using Pytorch
Stars: ✭ 43 (-90.95%)
Mutual labels:  nmt
Nmt List
A list of Neural MT implementations
Stars: ✭ 359 (-24.42%)
Mutual labels:  nmt
VNMT
Code for "Variational Neural Machine Translation" (EMNLP2016)
Stars: ✭ 54 (-88.63%)
Mutual labels:  nmt
Quality-Estimation1
机器翻译子任务-翻译质量评价-复现 WMT2018 阿里论文结果
Stars: ✭ 19 (-96%)
Mutual labels:  nmt
Seq2seq chatbot
基于seq2seq模型的简单对话系统的tf实现,具有embedding、attention、beam_search等功能,数据集是Cornell Movie Dialogs
Stars: ✭ 308 (-35.16%)
Mutual labels:  nmt
vat nmt
Implementation of "Effective Adversarial Regularization for Neural Machine Translation", ACL 2019
Stars: ✭ 22 (-95.37%)
Mutual labels:  nmt
Nmtpytorch
Sequence-to-Sequence Framework in PyTorch
Stars: ✭ 392 (-17.47%)
Mutual labels:  nmt
Nlp pytorch project
Embedding, NMT, Text_Classification, Text_Generation, NER etc.
Stars: ✭ 153 (-67.79%)
Mutual labels:  nmt
Variational-NMT
Variational Neural Machine Translation System
Stars: ✭ 37 (-92.21%)
Mutual labels:  nmt
Joeynmt
Minimalist NMT for educational purposes
Stars: ✭ 420 (-11.58%)
Mutual labels:  nmt
Neuralmonkey
An open-source tool for sequence learning in NLP built on TensorFlow.
Stars: ✭ 400 (-15.79%)
Mutual labels:  nmt
Ner Bert
BERT-NER (nert-bert) with google bert https://github.com/google-research.
Stars: ✭ 339 (-28.63%)
Mutual labels:  nmt

This is a fork of Fairseq(-py) with implementations of the following models:

Pervasive Attention - 2D Convolutional Neural Networks for Sequence-to-Sequence Prediction

An NMT models with two-dimensional convolutions to jointly encode the source and the target sequences.

Pervasive Attention also provides an extensive decoding grid that we leverage to efficiently train wait-k models.

See README.

Efficient Wait-k Models for Simultaneous Machine Translation

Transformer Wait-k models (Ma et al., 2019) with unidirectional encoders and with joint training of multiple wait-k paths.

See README.

Fairseq Requirements and Installation

  • PyTorch version >= 1.4.0
  • Python version >= 3.6
  • For training new models, you'll also need an NVIDIA GPU and NCCL

To install fairseq:

pip install fairseq

Installing from source

To install fairseq from source and develop locally:

git clone https://github.com/pytorch/fairseq
cd fairseq
pip install --editable .

License

fairseq(-py) is MIT-licensed. The license applies to the pre-trained models as well.

Citation

For Pervasive Attention, please cite:

@InProceedings{elbayad18conll,
    author ="Elbayad, Maha and Besacier, Laurent and Verbeek, Jakob",
    title = "Pervasive Attention: 2D Convolutional Neural Networks for Sequence-to-Sequence Prediction",
    booktitle = "Proceedings of the 22nd Conference on Computational Natural Language Learning",
    year = "2018",
 }

For our wait-k models, please cite:

@article{elbayad20waitk,
    title={Efficient Wait-k Models for Simultaneous Machine Translation},
    author={Elbayad, Maha and Besacier, Laurent and Verbeek, Jakob},
    journal={arXiv preprint arXiv:2005.08595},
    year={2020}
}

For Fairseq, please cite:

@inproceedings{ott2019fairseq,
  title = {fairseq: A Fast, Extensible Toolkit for Sequence Modeling},
  author = {Myle Ott and Sergey Edunov and Alexei Baevski and Angela Fan and Sam Gross and Nathan Ng and David Grangier and Michael Auli},
  booktitle = {Proceedings of NAACL-HLT 2019: Demonstrations},
  year = {2019},
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].