We describe our latest attempt at adaptive language modeling. At the heart of our approach is a Maximum Entropy (ME) model, which incorporates many knowledge sources in a consistent manner. The other components are a selective unigram cache, a conditional bigram cache, and a conventional static trigram. We describe the knowledge sources used to build such a model with ARPA's official WSJ corpus, and report on perplexity and word error rate results obtained with it. Then, three different adaptation paradigms are discussed, and an additional experiment, based on AP wire data, is used to compare them.
A Hybrid Approach to Adaptive Statistical Language Modeling
Published 1994 in Human Language Technology - The Baltic Perspectiv
ABSTRACT
PUBLICATION RECORD
- Publication year
1994
- Venue
Human Language Technology - The Baltic Perspectiv
- Publication date
1994-03-08
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-13 of 13 references · Page 1 of 1
CITED BY
Showing 1-34 of 34 citing papers · Page 1 of 1