The attentional mechanism has proven to be effective in improving end-to-end neural machine translation. However, due to the intricate structural divergence between natural languages, unidirectional attention-based models might only capture partial aspects of attentional regularities. We propose agreement-based joint training for bidirectional attention-based end-to-end neural machine translation. Instead of training source-to-target and target-to-source translation models independently, our approach encourages the two complementary models to agree on word alignment matrices on the same training data. Experiments on Chinese-English and English-French translation tasks show that agreement-based joint training significantly improves both alignment and translation quality over independent training.
Agreement-Based Joint Training for Bidirectional Attention-Based Neural Machine Translation
Yong Cheng,Shiqi Shen,Zhongjun He,W. He,Hua Wu,Maosong Sun,Yang Liu
Published 2015 in International Joint Conference on Artificial Intelligence
ABSTRACT
PUBLICATION RECORD
- Publication year
2015
- Venue
International Joint Conference on Artificial Intelligence
- Publication date
2015-12-15
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-20 of 20 references · Page 1 of 1
CITED BY
Showing 1-74 of 74 citing papers · Page 1 of 1