Neural Machine Translation

This page contains information about latest research on neural machine translation (NMT) at Stanford NLP group. We release our codebase which produces state-of-the-art results in various translation tasks such as English-German and English-Czech. In addtion, to encourage reproducibility and increase transparency, we release the preprocessed data that we used to train our models as well as our pretrained models that are readily usable with our codebase.

People

Codebase

For hybrid NMT, please use this codebase and cite:

Minh-Thang Luong, and Christopher D. Manning. 2016. Achieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models. In Association for Computational Linguistics (ACL). [pdf] [bib]

For general attention-based NMT, please use this codebase and cite:

Minh-Thang Luong, Hieu Pham, and Christopher D. Manning. 2015. Effective Approaches to Attention-based Neural Machine Translation. In Empirical Methods in Natural Language Processing (EMNLP). [pdf] [bib]

For pruning NMT, please cite (and let us know if you are interested in our code!):

Abigail See*, Minh-Thang Luong*, and Christopher D. Manning. 2016. Compression of Neural Machine Translation Models via Pruning. In Computational Natural Language Learning (CoNLL). [pdf] [bib]

Preprocessed Data

Pretrained Models

We release pretrained models that are readily usable with our Matlab code.
Note: to use these models, a GPU device is required. To convert these models to be used in a CPU, consider this script.

Contact Information

For any comments or questions, please email the first author.