All Projects → THUNLP-MT → Mt Reading List

THUNLP-MT / Mt Reading List

Licence: bsd-3-clause
A machine translation reading list maintained by Tsinghua Natural Language Processing Group

Programming Languages

TeX
3793 projects

Projects that are alternatives of or similar to Mt Reading List

Mt Paper Lists
MT paper lists (by conference)
Stars: ✭ 105 (-95.15%)
Mutual labels:  machine-translation
Awesome Ai Services
An overview of the AI-as-a-service landscape
Stars: ✭ 133 (-93.86%)
Mutual labels:  machine-translation
Deepl Translator
This module provides promised methods for translating text using DeepL Translator (https://www.deepl.com/translator) undocumented API.
Stars: ✭ 145 (-93.31%)
Mutual labels:  machine-translation
Reading Material
List of some useful blogs, books, courses, papers etc. 📚
Stars: ✭ 116 (-94.64%)
Mutual labels:  reading-list
Gtos
Code for AAAI2020 paper "Graph Transformer for Graph-to-Sequence Learning"
Stars: ✭ 129 (-94.04%)
Mutual labels:  machine-translation
Subword Nmt
Unsupervised Word Segmentation for Neural Machine Translation and Text Generation
Stars: ✭ 1,819 (-16.02%)
Mutual labels:  machine-translation
En Fr Mlt Tensorflow
English-French Machine Language Translation in Tensorflow
Stars: ✭ 99 (-95.43%)
Mutual labels:  machine-translation
Mtbook
《机器翻译:基础与模型》肖桐 朱靖波 著 - Machine Translation: Foundations and Models
Stars: ✭ 2,307 (+6.51%)
Mutual labels:  machine-translation
Recommends
A collection of resources and reading material that we recommend
Stars: ✭ 133 (-93.86%)
Mutual labels:  reading-list
Tensor2tensor
Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
Stars: ✭ 11,865 (+447.78%)
Mutual labels:  machine-translation
Nonautoreggenprogress
Tracking the progress in non-autoregressive generation (translation, transcription, etc.)
Stars: ✭ 118 (-94.55%)
Mutual labels:  machine-translation
Cluedatasetsearch
搜索所有中文NLP数据集,附常用英文NLP数据集
Stars: ✭ 2,112 (-2.49%)
Mutual labels:  machine-translation
Pytorch Dual Learning
Implementation of Dual Learning NMT on PyTorch
Stars: ✭ 141 (-93.49%)
Mutual labels:  machine-translation
Opus Mt
Open neural machine translation models and web services
Stars: ✭ 111 (-94.88%)
Mutual labels:  machine-translation
Deeply
PHP client for the DeepL.com translation API (unofficial)
Stars: ✭ 152 (-92.98%)
Mutual labels:  machine-translation
Effective Java Third Edition Reading Notes
Reading Effective Java, Third Edition in 11 days (3 hours per day) and make notes as much as possible.
Stars: ✭ 102 (-95.29%)
Mutual labels:  reading-list
Tokenizer
Fast and customizable text tokenization library with BPE and SentencePiece support
Stars: ✭ 132 (-93.91%)
Mutual labels:  machine-translation
Openkiwi
Open-Source Machine Translation Quality Estimation in PyTorch
Stars: ✭ 157 (-92.75%)
Mutual labels:  machine-translation
Nspm
🤖 Neural SPARQL Machines for Knowledge Graph Question Answering.
Stars: ✭ 156 (-92.8%)
Mutual labels:  machine-translation
Masakhane Mt
Machine Translation for Africa
Stars: ✭ 142 (-93.44%)
Mutual labels:  machine-translation

Machine Translation Reading List

This is a machine translation reading list maintained by the Tsinghua Natural Language Processing Group.

The past three decades have witnessed the rapid development of machine translation, especially for data-driven approaches such as statistical machine translation (SMT) and neural machine translation (NMT). Due to the dominance of NMT at the present time, priority is given to collecting important, up-to-date NMT papers; the Edinburgh/JHU MT research survey wiki has good coverage of older papers and a brief description for each sub-topic of MT. Our list is still incomplete and the categorization might be inappropriate. We will keep adding papers and improving the list. Any suggestions are welcome!

10 Must Reads

Tutorials and Surveys

Statistical Machine Translation

Word-based Models

Phrase-based Models

Syntax-based Models

Discriminative Training

System Combination

Human-centered SMT

Interactive SMT

Adaptation

Evaluation

Neural Machine Translation

Model Architecture

Attention Mechanism

Open Vocabulary

Training Objectives and Frameworks

Decoding

Low-resource Language Translation

Semi-supervised Learning

Unsupervised Learning

Pivot-based Methods

Data Augmentation Methods

Data Selection Methods

Transfer Learning

Meta Learning

Multilingual Machine Translation

Prior Knowledge Integration

Word/Phrase Constraints

Syntactic/Semantic Constraints

Coverage Constraints

Document-level Translation

Robustness

Interpretability

Linguistic Interpretation

Fairness and Diversity

Efficiency

Pre-Training

Non-Autoregressive Translation

Speech Translation and Simultaneous Translation

Multi-modality

Ensemble and Reranking

Domain Adaptation

Quality Estimation

Human-centered NMT

Interactive NMT

Automatic Post-Editing

Poetry Translation

Eco-friendly

Endangered Language Revitalization

Word Translation

WMT Winners

WMT is the most important annual international competition on machine translation. We collect the competition results on the news translation task since WMT 2016 (the First Conference of Machine Translation) and summarize the techniques used in the systems with the top performance. Currently, we focus on four directions: ZH-EN, EN-ZH, DE-EN, and EN-DE. The summarized algorithms might be incomplete; your suggestions are welcome!

WMT 2019

WMT 2018

  • The winner of ZH-EN: Tencent

    • System report: Mingxuan Wang, Li Gong, Wenhuan Zhu, Jun Xie, and Chao Bian. 2018. Tencent Neural Machine Translation Systems for WMT18. In Proceedings of the Third Conference on Machine Translation: Shared Task Papers.
    • Techniques: RNMT + Transformer + BPE + Rerank ensemble outputs with 48 features (including t2t R2l, t2t L2R, rnn L2R, rnn R2L etc.) + Back Translation + Joint Train with English to Chinese systems + Fine-tuning with selected data + Knowledge distillation
  • The winner of EN-ZH: GTCOM

    • System report: Chao Bei, Hao Zong, Yiming Wang, Baoyong Fan, Shiqi Li, and Conghu Yuan. 2018. An Empirical Study of Machine Translation for the Shared Task of WMT18. In Proceedings of the Third Conference on Machine Translation: Shared Task Papers.
    • Techniques: Transformer + Back-Translation + Data Filtering by rules, language models and translation models + BPE + Greedy Ensemble Decoding + Fine-Tuning with newstest2017 back translation
  • The winner of DE-EN: RWTH Aachen University

    • System report: Julian Schamper, Jan Rosendahl, Parnia Bahar, Yunsu Kim, Arne Nix, and Hermann Ney. 2018. The RWTH Aachen University Supervised Machine Translation Systems for WMT 2018. In Proceedings of the Third Conference on Machine Translation: Shared Task Papers.
    • Techniques: Ensemble of 3-strongest Transformer models + Data Selection + BPE + Fine-Tuning + Important Hyperparameters (batch size and model dimension)
  • The winner of EN-DE: Microsoft

WMT 2017

  • The winner of ZH-EN: Sogou

    • System report: Yuguang Wang, Shanbo Cheng, Liyang Jiang, Jiajun Yang, Wei Chen, Muze Li, Lin Shi, Yanfeng Wang, and Hongtao Yang. 2017. Sogou Neural Machine Translation Systems for WMT17. In Proceedings of the Second Conference on Machine Translation: Shared Task Papers.
    • Techniques: Encoder-Decoder with Attention + BPE + Reranking (R2L, T2S, N-gram language models) + Tagging Model + Name Entity Translation + Ensemble
  • The winner of EN-ZH, DE-EN and EN-DE: University of Edinburgh

    • System report: Rico Sennrich, Alexandra Birch, Anna Currey, Ulrich Germann, Barry Haddow, Kenneth Heafield, Antonio Valerio Miceli Barone, and Philip Williams. 2017. The University of Edinburgh’s Neural MT Systems for WMT17. In Proceedings of the Second Conference on Machine Translation: Shared Task Papers.
    • Techniques: Encoder-Decoder with Attention + Deep Model + Layer Normalization + Weight Tying + Back-Translation + BPE + Reranking(L2R, R2L) + Ensemble

WMT 2016

  • The winner of DE-EN: University of Regensburg

    • System report: Failed to find it
    • Techniques: Failed to find it
  • The winner of EN-DE: University of Edinburgh

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].