We study the properties of the MDL (or maximum penalized complexity) estimator for Regression and Classification, where the underlying model class is countable. We show in particular a finite bound on the Hellinger losses under the only assumption that there is a “true” model contained in the class. This implies almost sure convergence of the predictive distribution to the true one at a fast rate. It corresponds to Solomono’s central theorem of universal induction, however with a bound that is exponentially larger.
Strong Asymptotic Assertions for Discrete MDL in Regression and Classification
Published 2005 in arXiv.org
ABSTRACT
PUBLICATION RECORD
- Publication year
2005
- Venue
arXiv.org
- Publication date
2005-01-31
- Fields of study
Mathematics, Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-14 of 14 references · Page 1 of 1
CITED BY
Showing 1-7 of 7 citing papers · Page 1 of 1