Decoding with Large-Scale Neural Language Models Improves Translation

Ashish Vaswani,Yinggong Zhao,Victoria Fossum,David Chiang

Published 2013 in Conference on Empirical Methods in Natural Language Processing

ABSTRACT

We explore the application of neural language models to machine translation. We develop a new model that combines the neural probabilistic language model of Bengio et al., rectified linear units, and noise-contrastive estimation, and we incorporate it into a machine translation system both by reranking k-best lists and by direct integration into the decoder. Our large-scale, large-vocabulary experiments across four language pairs show that our neural language model improves translation quality by up to 1.1 Bleu.

PUBLICATION RECORD

  • Publication year

    2013

  • Venue

    Conference on Empirical Methods in Natural Language Processing

  • Publication date

    2013-10-01

  • Fields of study

    Linguistics, Computer Science

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-25 of 25 references · Page 1 of 1

CITED BY

Showing 1-100 of 277 citing papers · Page 1 of 3