MDL Convergence Speed for Bernoulli Sequences

Mathematics – Statistics Theory

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

28 pages

Scientific paper

10.1007/s11222-006-6746-3

The Minimum Description Length principle for online sequence estimation/prediction in a proper learning setup is studied. If the underlying model class is discrete, then the total expected square loss is a particularly interesting performance measure: (a) this quantity is finitely bounded, implying convergence with probability one, and (b) it additionally specifies the convergence speed. For MDL, in general one can only have loss bounds which are finite but exponentially larger than those for Bayes mixtures. We show that this is even the case if the model class contains only Bernoulli distributions. We derive a new upper bound on the prediction error for countable Bernoulli classes. This implies a small bound (comparable to the one for Bayes mixtures) for certain important model classes. We discuss the application to Machine Learning tasks such as classification and hypothesis testing, and generalization to countable classes of i.i.d. models.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

MDL Convergence Speed for Bernoulli Sequences does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with MDL Convergence Speed for Bernoulli Sequences, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and MDL Convergence Speed for Bernoulli Sequences will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-89196

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.