We propose multi-way, multilingual neural machine translation. The proposed approach enables a single neural translation model to translate between multiple languages, with a number of parameters that grows only linearly with the number of languages. This is made possible by having a single attention mechanism that is shared across all language pairs. We train the proposed multi-way, multilingual model on ten language pairs from WMT'15 simultaneously and observe clear performance improvements over models trained on only one language pair. In particular, we observe that the proposed model significantly improves the translation quality of low-resource language pairs.
Multi-Way, Multilingual Neural Machine Translation with a Shared Attention Mechanism
Orhan Firat,Kyunghyun Cho,Yoshua Bengio
Published 2016 in North American Chapter of the Association for Computational Linguistics
ABSTRACT
PUBLICATION RECORD
- Publication year
2016
- Venue
North American Chapter of the Association for Computational Linguistics
- Publication date
2016-01-06
- Fields of study
Mathematics, Linguistics, Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-25 of 25 references · Page 1 of 1