We present a simple, but surprisingly effective, method of self-training a two-phase parser-reranker system using readily available unlabeled data. We show that this type of bootstrapping is possible for parsing when the bootstrapped parses are processed by a discriminative reranker. Our improved model achieves an f-score of 92.1%, an absolute 1.1% improvement (12% error reduction) over the previous best result for Wall Street Journal parsing. Finally, we provide some analysis to better understand the phenomenon.
Effective Self-Training for Parsing
David McClosky,Eugene Charniak,Mark Johnson
Published 2006 in North American Chapter of the Association for Computational Linguistics
ABSTRACT
PUBLICATION RECORD
- Publication year
2006
- Venue
North American Chapter of the Association for Computational Linguistics
- Publication date
2006-06-04
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-19 of 19 references · Page 1 of 1