Strong asymptotic assertions for discrete MDL in regression and classification
Date
Authors
Poland, Jan
Hutter, Marcus
Journal Title
Journal ISSN
Volume Title
Publisher
University of Twente
Abstract
We study the properties of the MDL (or maximum penalized
complexity) estimator for Regression and Classification, where the
underlying model class is countable. We show in particular a
finite bound on the Hellinger losses under the only assumption
that there is a ``true'' model contained in the class. This implies
almost sure convergence of the predictive distribution to the true
one at a fast rate. It corresponds to Solomonoff's central theorem
of universal induction, however with a bound that is exponentially
larger.
Description
Citation
Collections
Source
Type
Book Title
Proceedings of the 14th Dutch-Belgium Conference on Machine Learning Benelearn'05