All Projects → deep-diver → En Fr Mlt Tensorflow

deep-diver / En Fr Mlt Tensorflow

English-French Machine Language Translation in Tensorflow

Projects that are alternatives of or similar to En Fr Mlt Tensorflow

Texar Pytorch
Integrating the Best of TF into PyTorch, for Machine Learning, Natural Language Processing, and Text Generation. This is part of the CASL project: http://casl-project.ai/
Stars: ✭ 636 (+542.42%)
Mutual labels:  machine-translation
Mtnt
Code for the collection and analysis of the MTNT dataset
Stars: ✭ 48 (-51.52%)
Mutual labels:  machine-translation
Chinesenlp
Datasets, SOTA results of every fields of Chinese NLP
Stars: ✭ 1,206 (+1118.18%)
Mutual labels:  machine-translation
Nlg Eval
Evaluation code for various unsupervised automated metrics for Natural Language Generation.
Stars: ✭ 822 (+730.3%)
Mutual labels:  machine-translation
Sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on Apache MXNet
Stars: ✭ 990 (+900%)
Mutual labels:  machine-translation
Fasttext multilingual
Multilingual word vectors in 78 languages
Stars: ✭ 1,067 (+977.78%)
Mutual labels:  machine-translation
Seq2seq
Minimal Seq2Seq model with Attention for Neural Machine Translation in PyTorch
Stars: ✭ 552 (+457.58%)
Mutual labels:  machine-translation
Niutrans.smt
NiuTrans.SMT is an open-source statistical machine translation system developed by a joint team from NLP Lab. at Northeastern University and the NiuTrans Team. The NiuTrans system is fully developed in C++ language. So it runs fast and uses less memory. Currently it supports phrase-based, hierarchical phrase-based and syntax-based (string-to-tree, tree-to-string and tree-to-tree) models for research-oriented studies.
Stars: ✭ 90 (-9.09%)
Mutual labels:  machine-translation
Bartycrouch
Localization/I18n: Incrementally update/translate your Strings files from .swift, .h, .m(m), .storyboard or .xib files.
Stars: ✭ 1,032 (+942.42%)
Mutual labels:  machine-translation
Udacity Natural Language Processing Nanodegree
Tutorials and my solutions to the Udacity NLP Nanodegree
Stars: ✭ 73 (-26.26%)
Mutual labels:  machine-translation
Turbotransformers
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
Stars: ✭ 826 (+734.34%)
Mutual labels:  machine-translation
Witwicky
Witwicky: An implementation of Transformer in PyTorch.
Stars: ✭ 21 (-78.79%)
Mutual labels:  machine-translation
Thot
Thot toolkit for statistical machine translation
Stars: ✭ 53 (-46.46%)
Mutual labels:  machine-translation
Nematus
Open-Source Neural Machine Translation in Tensorflow
Stars: ✭ 730 (+637.37%)
Mutual labels:  machine-translation
Transformers without tears
Transformers without Tears: Improving the Normalization of Self-Attention
Stars: ✭ 80 (-19.19%)
Mutual labels:  machine-translation
Seq2seq
A general-purpose encoder-decoder framework for Tensorflow
Stars: ✭ 5,455 (+5410.1%)
Mutual labels:  machine-translation
Machine Translation
Stars: ✭ 51 (-48.48%)
Mutual labels:  machine-translation
Deep Learning Drizzle
Drench yourself in Deep Learning, Reinforcement Learning, Machine Learning, Computer Vision, and NLP by learning from these exciting lectures!!
Stars: ✭ 9,717 (+9715.15%)
Mutual labels:  machine-translation
Opennmt Tf
Neural machine translation and sequence learning using TensorFlow
Stars: ✭ 1,223 (+1135.35%)
Mutual labels:  machine-translation
Comet
A Neural Framework for MT Evaluation
Stars: ✭ 58 (-41.41%)
Mutual labels:  machine-translation

MLT (EN to FR ) TensorFlow

In this project, I am going to build language translation model called seq2seq model or encoder-decoder model in TensorFlow. The objective of the model is translating English sentences to French sentences. I am going to show the detailed steps, and they will answer to the questions like how to preprocess the dataset, how to define inputs, how to define encoder model, how to define decoder model, how to build the entire seq2seq model, how to calculate the loss and clip gradients, and how to train and get prediction. Please open the IPython notebook file to see the full workflow and detailed descriptions.

This is a part of Udacity's Deep Learning Nanodegree. Some codes/functions (save, load, measuring accuracy, etc) are provided by Udacity. However, majority part is implemented by myself along with much richer explanations and references on each section.

You can find only the model part explained in my medium post. https://medium.com/@parkchansung/seq2seq-model-in-tensorflow-ec0c557e560f

Brief Overview of the Contents

Data preprocessing

In this section, you will see how to get the data, how to create lookup table, and how to convert raw text to index based array with the lookup table.

Drawing

Build model

In short, this section will show how to define the Seq2Seq model in TensorFlow. The below steps (implementation) will be covered.

  • (1) define input parameters to the encoder model
    • enc_dec_model_inputs
  • (2) build encoder model
    • encoding_layer
  • (3) define input parameters to the decoder model
    • enc_dec_model_inputs, process_decoder_input, decoding_layer
  • (4) build decoder model for training
    • decoding_layer_train
  • (5) build decoder model for inference
    • decoding_layer_infer
  • (6) put (4) and (5) together
    • decoding_layer
  • (7) connect encoder and decoder models
    • seq2seq_model
  • (8) train and estimate loss and accuracy

Training

This section is about putting previously defined functions together to build an actual instance of the model. Furthermore, it will show how to define cost function, how to apply optimizer to the cost function, and how to modify the value of the gradients in the TensorFlow's optimizer module to perform gradient clipping.

Prediction

Nothing special but showing the prediction result.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].