All Projects → soskek → convolutional_seq2seq

soskek / convolutional_seq2seq

Licence: BSD-3-Clause license
fairseq: Convolutional Sequence to Sequence Learning (Gehring et al. 2017) by Chainer

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to convolutional seq2seq

Deeplearningmugenknock
でぃーぷらーにんぐを無限にやってディープラーニングでDeepLearningするための実装CheatSheet
Stars: ✭ 684 (+985.71%)
Mutual labels:  chainer, seq2seq
MTA-LSTM-TensorFlow
TensorFlow reimplementation of Topic-to-Essay Generation with Neural Networks.
Stars: ✭ 67 (+6.35%)
Mutual labels:  seq2seq
Seq2Seq-chatbot
TensorFlow Implementation of Twitter Chatbot
Stars: ✭ 18 (-71.43%)
Mutual labels:  seq2seq
Visual-Attention-Model
Chainer implementation of Deepmind's Visual Attention Model paper
Stars: ✭ 27 (-57.14%)
Mutual labels:  chainer
chainer-grad-cam
Chainer implementation of Grad-CAM
Stars: ✭ 20 (-68.25%)
Mutual labels:  chainer
Transformer Temporal Tagger
Code and data form the paper BERT Got a Date: Introducing Transformers to Temporal Tagging
Stars: ✭ 55 (-12.7%)
Mutual labels:  seq2seq
Naver-AI-Hackathon-Speech
2019 Clova AI Hackathon : Speech - Rank 12 / Team Kai.Lib
Stars: ✭ 26 (-58.73%)
Mutual labels:  seq2seq
avsr-tf1
Audio-Visual Speech Recognition using Sequence to Sequence Models
Stars: ✭ 76 (+20.63%)
Mutual labels:  seq2seq
NeuralCodeTranslator
Neural Code Translator provides instructions, datasets, and a deep learning infrastructure (based on seq2seq) that aims at learning code transformations
Stars: ✭ 32 (-49.21%)
Mutual labels:  seq2seq
seq2seq-autoencoder
Theano implementation of Sequence-to-Sequence Autoencoder
Stars: ✭ 12 (-80.95%)
Mutual labels:  seq2seq
ChainerPruner
ChainerPruner: Channel Pruning framework for Chainer
Stars: ✭ 21 (-66.67%)
Mutual labels:  chainer
chinese ancient poetry
seq2seq attention tensorflow textrank context
Stars: ✭ 30 (-52.38%)
Mutual labels:  seq2seq
chainer2pytorch
Converts Chainer modules to PyTorch, parameters included.
Stars: ✭ 36 (-42.86%)
Mutual labels:  chainer
TextSumma
reimplementing Neural Summarization by Extracting Sentences and Words
Stars: ✭ 16 (-74.6%)
Mutual labels:  seq2seq
BMI219-2017-ProteinFolding
UCSF BMI219 Deep Learning (2017), Coding example (Prediction of protein folding with RNN and CNN)
Stars: ✭ 14 (-77.78%)
Mutual labels:  chainer
pyner
🌈 Implementation of Neural Network based Named Entity Recognizer (Lample+, 2016) using Chainer.
Stars: ✭ 45 (-28.57%)
Mutual labels:  chainer
3dgan-chainer
📦 A Chainer implementation of 3D Generative Adversarial Network.
Stars: ✭ 25 (-60.32%)
Mutual labels:  chainer
beam search
Beam search for neural network sequence to sequence (encoder-decoder) models.
Stars: ✭ 31 (-50.79%)
Mutual labels:  seq2seq
parse seq2seq
A tensorflow implementation of neural sequence-to-sequence parser for converting natural language queries to logical form.
Stars: ✭ 26 (-58.73%)
Mutual labels:  seq2seq
deep-learning-tutorial-with-chainer
Deep learning tutorial with Chainer
Stars: ✭ 25 (-60.32%)
Mutual labels:  chainer

Convolutional Sequence to Sequence Learning

Chainer-based Python implementation of a convolutional seq2seq model.

This is derived from Chainer's official seq2seq example.

See Convolutional Sequence to Sequence Learning, Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, Yann N. Dauphin, arxiv, 2017. blog post, Torch code.

Requirement

  • Python 3.6.0+
  • Chainer 2.0.0+ (this version is strictly required)
  • numpy 1.12.1+
  • cupy 1.0.0+ (if using gpu)
  • and their dependencies

Prepare Dataset

You can use any parallel corpus.
For example, run download_wmt.sh which downloads and decompresses training dataset and development dataset from WMT/europal into your current directory. These files and their paths are set in training script seq2seq.py as default.

How to Run

PYTHONIOENCODING=utf-8 python -u seq2seq.py -g=0 -i DATA_DIR -o SAVE_DIR

During training, logs for loss, perplexity, word accuracy and time are printed at a certain internval, in addition to validation tests (perplexity and BLEU for generation) every half epoch. And also, generation test is performed and printed for checking training progress.

Arguments

  • -g: your gpu id. If cpu, set -1.
  • -i DATA_DIR, -s SOURCE, -t TARGET, -svalid SVALID, -tvalid TVALID:
    DATA_DIR directory needs to include a pair of training dataset SOURCE and TARGET with a pair of validation dataset SVALID and TVALID. Each pair should be parallell corpus with line-by-line sentence alignment.
  • -o SAVE_DIR: JSON log report file and a model snapshot will be saved in SAVE_DIR directory (if it does not exist, it will be automatically made).
  • -e: max epochs of training corpus.
  • -b: minibatch size.
  • -u: size of units and word embeddings.
  • -l: number of layers in both the encoder and the decoder.
  • --source-vocab: max size of vocabulary set of source language
  • --target-vocab: max size of vocabulary set of target language
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].