All Projects → bzhangGo → zero

bzhangGo / zero

Licence: BSD-3-Clause license
Zero -- A neural machine translation system

Programming Languages

python
139335 projects - #7 most used programming language
perl
6916 projects

Projects that are alternatives of or similar to zero

transformer
Neutron: A pytorch based implementation of Transformer and its variants.
Stars: ✭ 60 (-50.41%)
Mutual labels:  transformer, neural-machine-translation, average-attention-network
NiuTrans.NMT
A Fast Neural Machine Translation System. It is developed in C++ and resorts to NiuTensor for fast tensor APIs.
Stars: ✭ 112 (-7.44%)
Mutual labels:  transformer, neural-machine-translation
Joeynmt
Minimalist NMT for educational purposes
Stars: ✭ 420 (+247.11%)
Mutual labels:  transformer, neural-machine-translation
transformer-slt
Sign Language Translation with Transformers (COLING'2020, ECCV'20 SLRTP Workshop)
Stars: ✭ 92 (-23.97%)
Mutual labels:  transformer, neural-machine-translation
Pytorch Seq2seq
Tutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.
Stars: ✭ 3,418 (+2724.79%)
Mutual labels:  transformer, neural-machine-translation
Nmt Keras
Neural Machine Translation with Keras
Stars: ✭ 501 (+314.05%)
Mutual labels:  transformer, neural-machine-translation
Sockeye
Sequence-to-sequence framework with a focus on Neural Machine Translation based on Apache MXNet
Stars: ✭ 990 (+718.18%)
Mutual labels:  transformer, neural-machine-translation
Transformer Dynet
An Implementation of Transformer (Attention Is All You Need) in DyNet
Stars: ✭ 57 (-52.89%)
Mutual labels:  transformer, neural-machine-translation
Njunmt Tf
An open-source neural machine translation system developed by Natural Language Processing Group, Nanjing University.
Stars: ✭ 97 (-19.83%)
Mutual labels:  transformer, neural-machine-translation
Neural-Machine-Translation
Several basic neural machine translation models implemented by PyTorch & TensorFlow
Stars: ✭ 29 (-76.03%)
Mutual labels:  transformer, neural-machine-translation
NLP-paper
🎨 🎨NLP 自然语言处理教程 🎨🎨 https://dataxujing.github.io/NLP-paper/
Stars: ✭ 23 (-80.99%)
Mutual labels:  transformer
sentencepiece-jni
Java JNI wrapper for SentencePiece: unsupervised text tokenizer for Neural Network-based text generation.
Stars: ✭ 26 (-78.51%)
Mutual labels:  neural-machine-translation
AdaSpeech
AdaSpeech: Adaptive Text to Speech for Custom Voice
Stars: ✭ 108 (-10.74%)
Mutual labels:  transformer
MusicTransformer-Pytorch
MusicTransformer written for MaestroV2 using the Pytorch framework for music generation
Stars: ✭ 106 (-12.4%)
Mutual labels:  transformer
SegFormer
Official PyTorch implementation of SegFormer
Stars: ✭ 1,264 (+944.63%)
Mutual labels:  transformer
dynmt-py
Neural machine translation implementation using dynet's python bindings
Stars: ✭ 17 (-85.95%)
Mutual labels:  neural-machine-translation
CrabNet
Predict materials properties using only the composition information!
Stars: ✭ 57 (-52.89%)
Mutual labels:  transformer
MT-Preparation
Machine Translation (MT) Preparation Scripts
Stars: ✭ 15 (-87.6%)
Mutual labels:  neural-machine-translation
semantic-segmentation
SOTA Semantic Segmentation Models in PyTorch
Stars: ✭ 464 (+283.47%)
Mutual labels:  transformer
wenet
Production First and Production Ready End-to-End Speech Recognition Toolkit
Stars: ✭ 2,384 (+1870.25%)
Mutual labels:  transformer

Zero

A neural machine translation system implemented by python2 + tensorflow.

Features

  1. Multi-Process Data Loading/Processing (Problems Exist)
  2. Multi-GPU Training/Decoding
  3. Gradient Aggregation

Papers

We associate each paper below with a readme file link. Please click the paper link you are interested for more details.

Supported Models

Requirements

  • python2.7
  • tensorflow <= 1.13.2

Usage

How to use this toolkit for machine translation?

TODO:

  1. organize the parameters and interpretations in config.
  2. reformat and fulfill code comments
  3. simplify and remove unecessary coding
  4. improve rnn models

Citation

If you use the source code, please consider citing the follow paper:

@InProceedings{D18-1459,
  author = 	"Zhang, Biao
		and Xiong, Deyi
		and su, jinsong
		and Lin, Qian
		and Zhang, Huiji",
  title = 	"Simplifying Neural Machine Translation with Addition-Subtraction Twin-Gated Recurrent Networks",
  booktitle = 	"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
  year = 	"2018",
  publisher = 	"Association for Computational Linguistics",
  pages = 	"4273--4283",
  location = 	"Brussels, Belgium",
  url = 	"http://aclweb.org/anthology/D18-1459"
}

If you are interested in the CAEncoder model, please consider citing our TASLP paper:

@article{Zhang:2017:CRE:3180104.3180106,
 author = {Zhang, Biao and Xiong, Deyi and Su, Jinsong and Duan, Hong},
 title = {A Context-Aware Recurrent Encoder for Neural Machine Translation},
 journal = {IEEE/ACM Trans. Audio, Speech and Lang. Proc.},
 issue_date = {December 2017},
 volume = {25},
 number = {12},
 month = dec,
 year = {2017},
 issn = {2329-9290},
 pages = {2424--2432},
 numpages = {9},
 url = {https://doi.org/10.1109/TASLP.2017.2751420},
 doi = {10.1109/TASLP.2017.2751420},
 acmid = {3180106},
 publisher = {IEEE Press},
 address = {Piscataway, NJ, USA},
}

Reference

When developing this repository, I referred to the following projects:

Contact

For any questions or suggestions, please feel free to contact Biao Zhang

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].