Open Research will be unavailable from 10.15am - 11am on Saturday 14th March 2026 AEDT due to scheduled maintenance.
 

Asymptotics of discrete (MDL) for online prediction

dc.contributor.authorPoland, Jan
dc.contributor.authorHutter, Marcus
dc.date.accessioned2015-12-10T22:40:57Z
dc.date.issued2005
dc.date.updated2016-02-24T11:44:49Z
dc.description.abstractMinimum description length (MDL) is an important principle for induction and prediction, with strong relations to optimal Bayesian learning. This paper deals with learning processes which are independent and identically distributed (i.i.d.) by means of two-part MDL, where the underlying model class is countable. We consider the online learning framework, i.e., observations come in one by one, and the predictor is allowed to update its state of mind after each time step. We identify two ways of predicting by MDL for this setup, namely, a static and a dynamic one. (A third variant, hybrid MDL, will turn out inferior.) We will prove that under the only assumption that the data is generated by a distribution contained in the model class, the MDL predictions converge to the true values almost surely. This is accomplished by proving finite bounds on the quadratic, the Hellinger, and the Kullback-Leibler loss of the MDL learner, which are, however, exponentially worse than for Bayesian prediction. We demonstrate that these bounds are sharp, even for model classes containing only Bernoulli distributions. We show how these bounds imply regret bounds for arbitrary loss functions. Our results apply to a wide range of setups, namely, sequence prediction, pattern classification, regression, and universal induction in the sense of algorithmic information theory among others.
dc.identifier.issn0018-9448
dc.identifier.urihttp://hdl.handle.net/1885/57671
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE Inc)
dc.rightsCopyright Information: http://www.sherpa.ac.uk/romeo/issn/0018-9448/..."Author's post-print on Author's server or Institutional server" from SHERPA/RoMEO site (as at 31/08/15).;© 2005 IEEE. Personal use of this material is permitted. Permission from IEEE
dc.sourceIEEE Transactions on Information Theory
dc.subjectKeywords: Learning systems; Mathematical models; Pattern recognition; Probability distributions; Regression analysis; Theorem proving; Algorithmic information theory; Classification consistency; Discrete model class; Minimum description length (MDL); Sequence predi Algorithmic information theory; Classification; Consistency; Discrete model class; Loss bounds; Minimum description length (MDL); Regression; Sequence prediction; Stabilization; Universal induction
dc.titleAsymptotics of discrete (MDL) for online prediction
dc.typeJournal article
local.bibliographicCitation.issue11
local.bibliographicCitation.lastpage3795
local.bibliographicCitation.startpage3780
local.contributor.affiliationPoland, Jan, Hokkaido University
local.contributor.affiliationHutter, Marcus, College of Engineering and Computer Science, ANU
local.contributor.authoruidHutter, Marcus, u4350841
local.description.notesImported from ARIES
local.identifier.absfor080109 - Pattern Recognition and Data Mining
local.identifier.ariespublicationu8803936xPUB410
local.identifier.citationvolume51
local.identifier.doi10.1109/TIT.2005.856956
local.identifier.scopusID2-s2.0-27744462709
local.type.statusPublished Version

Downloads

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
01_Poland_Asymptotics_of_discrete_(MDL)_2005.pdf
Size:
499.18 KB
Format:
Adobe Portable Document Format