Statistical Learning of Arbitrary Computable Classifiers

Computer Science – Learning

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

Expanded the section on prior work and added references

Scientific paper

Statistical learning theory chiefly studies restricted hypothesis classes, particularly those with finite Vapnik-Chervonenkis (VC) dimension. The fundamental quantity of interest is the sample complexity: the number of samples required to learn to a specified level of accuracy. Here we consider learning over the set of all computable labeling functions. Since the VC-dimension is infinite and a priori (uniform) bounds on the number of samples are impossible, we let the learning algorithm decide when it has seen sufficient samples to have learned. We first show that learning in this setting is indeed possible, and develop a learning algorithm. We then show, however, that bounding sample complexity independently of the distribution is impossible. Notably, this impossibility is entirely due to the requirement that the learning algorithm be computable, and not due to the statistical nature of the problem.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Statistical Learning of Arbitrary Computable Classifiers does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Statistical Learning of Arbitrary Computable Classifiers, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Statistical Learning of Arbitrary Computable Classifiers will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-620712

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.