- Oct 2014 — I had a fruitful summer internship at Google hosted by Quoc Le. Together with other great members in Google Brain team, Ilya Sutskever, Oriol Vinyals, Wojciech Zaremba, we built a neural machine translation system that, for the first time, outperforms the state-of-the-art system in the English-to-French WMT'14 translation task. See our ACL'15 paper.
- Jul 2013 — Our TACL'13 paper "Parsing entire discourses as very long strings: capturing topic continuity in grounded language learning" with Prof. Mark Johnson and Prof. Michael C. Frank is now available. This work is done in Fall, 2012 at Macquarie University, Australia.
- Jun 2013 — Our CoNLL'13 paper "Better Word Representations with Recursive Neural Networks for Morphology" with Richard Socher and Prof. Christopher Manning is now available. See you in Sofia!
- Apr 2012 — This Winter quarter, I am officially aligned with Prof. Christopher Manning, i.e., my advisor!
- Jan 2012 — Happy new year, everyone! This Spring quarter, I am back to Gates, joining Prof. Christopher Manning's NLP group as part of the research rotation program.
- Sep 2011 — This Fall quarter, I am having a research rotation with Prof. Noah D. Goodman, exploring computational models for human reading time prediction.
- Aug 2011 — Paper accepted at ASRU'2011 "A Trajectory-based Parallel Model Combination with a Unified Static and Dynamic Parameter Compensation For Noisy Speech Recognition" with Prof. Khe Chai Sim.
- Apr 2011 — I will be joining the Stanford CS department this Fall as a PhD candidate (Sep 2011).
I am from Vietnam and in case you haven't heard about it, my country has recently discovered Son Doong, the world's largest cave!
I'll be graduating in 2016! Here is my (often outdated) CV.
- Dec 2015 — In Da Nang, Vietnam, for a panel discussion on "New trends in Spoken Language Translation" at IWSLT. Exciting! We also have a winning entry for English-German TED talk translation, 26% error reduction from the 2nd place. See here.
- Nov 2015 — Guest lecture on Neural Machine Translation in the Stanford NLP class CS224N.
- Oct 2015 — Visit IBM Watson Research, New York, and give a talk on Neural Machine Translation.
- Sep 2015 — We have three papers published, two at EMNLP and one at CogACLL, in Lisbon, Portugal. Our paper "Effective Approaches to Attention-based Neural Machine Translation" achieves SOTA results in English-German translation. We have released our Matlab code too!
- Jul 2015 — We have three papers published, two at ACL and one at CoNLL, in Beijing, China. Our paper "Addressing the Rare Word Problem in Neural Machine Translation" describes the very first end-to-end neural machine translation that achieves state-of-the-art performance in English-French translation.
- May 2015 — We have two papers at the NAACL VSM workshop, in Colorado, USA. We achieve state-of-the-art performances in the cross-lingual document classification task and code is released too!
I am now mainly into deep learning models for NLP tasks, specifically machine translation. Occassionaly, I build parsers used for psycholinguistics applications :)
Before Stanford: I worked in many scholarly Digital Library projects – document structure discovery and keyphrase extraction. I also investigated in model-based techniques for robust Speech Recognition systems.
- Minh-Thang Luong and Christopher D. Manning. Stanford Neural Machine Translation Systems for Spoken Language Domain. IWSLT’15. Da Nang, Vietnam. Oral.
[ Paper ] [ Bib ] [ Slides ] [ IWSLT result overview ] [ SOTA English-German system for TED talks, 26% better than the 2nd place. ]
- Minh-Thang Luong, Hieu Pham, and Christopher D. Manning. Effective Approaches to Attention-based Neural Machine Translation. EMNLP’15. Lisbon, Portugal. Oral.
[ Paper ] [ Bib ] [ Slides ] [ Project page ] [ SOTA results in WMT English-German translation ]
- Jiwei Li, Minh-Thang Luong, Dan Jurafsky, and Eduard Hovy. When Are Tree Structures Necessary for Deep Learning of Representations?. EMNLP’15. Lisbon, Portugal. Oral.
[ Paper ] [ Bib ]
- Minh-Thang Luong, Timothy J. O'Donnell, and Noah D. Goodman. Evaluating Models of Computation and Storage in Human Sentence Processing. CogACLL’15. Lisbon, Portugal. Oral.
[ Paper ] [ Bib ] [ Slides ] [ Code (Earleyx parser) ]
- Minh-Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, Wojciech Zaremba. Addressing the Rare Word Problem in Neural Machine Translation. ACL’15. Beijing, China. Oral.
[ Paper ] [ Bib ] [ Slides ] [ SOTA results in WMT English-French translation ]
- Jiwei Li, Minh-Thang Luong, Dan Jurafsky. A Hierarchical Neural Autoencoder for Paragraphs and Documents. ACL’15. Beijing, China. Oral.
[ Paper ]
- Minh-Thang Luong, Michael Kayser, and Christopher D. Manning. Deep Neural Language Models for Machine Translation. CoNLL’15. Beijing China. Short paper, poster.
[ Paper ] [ Bib ] [ Poster ] [ Project page for code. ]
- Minh-Thang Luong, Hieu Pham, and Christopher D. Manning. Bilingual Word Representations with Monolingual Quality in Mind. NAACL’15 VSM workshop. Denver, USA.
[ Paper ] [ Bib ] [ Project page for code and trained embeddings. ]
- Hieu Pham, Minh-Thang Luong, Christopher D. Manning. Learning Distributed Representations for Multilingual Text Sequences. NAACL’15 VSM workshop. Denver, USA.
[ Paper ]
- Minh-Thang Luong, Richard Socher, and Christopher D. Manning. 2013. Better Word Representations with Recursive Neural Networks for Morphology. Conference on Computational Natural Language Learning (CoNLL’13). Sofia, Bulgaria.
[ Paper ] [ Bib ] [ Dataset ] [ Project page for word vectors and other information. ]
- Minh-Thang Luong, Michael C. Frank, and Mark Johnson. 2013. Parsing entire discourses as very long strings: Capturing topic continuity in grounded language learning. Transactions of the Association for Computational Linguistics (TACL’13).
[ Paper ] [ Bib ] [ Code (Earleyx parser) ]
Before Stanford: 2011
- Khe Chai Sim and Minh-Thang Luong. 2011. A Trajectory-based Parallel Model Combination with a Unified Static and Dynamic Parameter Compensation For Noisy Speech Recognition. Automatic Speech Recognition and Understanding (ASRU’11). Hawaii, USA. Oral presentation.
[ Paper ] [ Poster (.pdf) ]
- Minh-Thang Luong, Thuy Dung Nguyen, and Min-Yen Kan. 2011. Logical Structure Recovery in Scholarly Articles with Rich Document Features. International Journal of Digital Library Systems (IJDLS) 1(4). pp 1-23. Invited paper.
[ Paper ] [ doi:10.4018/jdls.2010100101 ]
- Minh-Thang Luong, Preslav Nakov and Min-Yen Kan. 2010. A Hybrid Morpheme-Word Representation for Machine Translation of Morphologically Rich Languages. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’10). Boston, Massachusetts, USA. pp.148-157. Oral presentation.
[ Paper ] [ ACL Anthology (D10-1015) ]
- Minh-Thang Luong and Min-Yen Kan. 2010. Enhancing Morphological Alignment for Translating Highly Inflected Languages. In Proceedings of the International Conference on Computational Linguistics (COLING’10). Beijing, China. pp. 743-751. Oral presentation.
[ Paper ] [ ACL Anthology (C10-1084) ]
- Thuy Dung Nguyen and Minh-Thang Luong. 2010. WINGNUS: Keyphrase Extraction Utilizing Document Logical Structure. In Proceedings of the ACL 2010 Workshop on Evaluation Exercises on Semantic Evaluation (SemEval-2). pp. 166-169. Poster presentation.
Rank 3rd in the task #5 – Automatic Keyphrase Extraction from Scientific Articles.
[ Paper ] [ ACL Anthology (S10-1035) ]
- Thuy Dung Nguyen, Min-Yen Kan, Dinh-Trung Dang, Markus Hänse, Ching Hoi Andy Hong, Minh-Thang Luong, Jesse Prabawa Gozali, Kazunari Sugiyama and Yee Fan Tan (2010). ForeCite: Towards a Reader-centric Scholarly Digital Library. In Proceedings of the Joint Conference on Digital Libraries (JCDL’10). Poster Paper (1 page).
[ Paper ] [ Poster (.png) ] [ doi:10.1145/1816123.1816193 ]
SoftwareBelow are some links to the demos of software written by myself or with members in WING group:
- ParsCit: An open-source CRF Reference String and Logical Document Structure Parsing Package
- GSNAP: Google Scholar - Network Analysis Package
- ForeCite: A Reader-centric Digital Library
NotesThis is my latest notes, containing all tips/commands/notes about various tools and languages. They are arranged in structures so that I could write a script later to split into separate files, one per language or tool ... For now, sorry for such a mess ...
Below are my notes collected randomly while I were learning some new tools and stuffs. To be organized:) :
See more in my academic page!