Statistics – Machine Learning
Scientific paper
2010-11-23
Statistics
Machine Learning
Scientific paper
In this article, we derive concentration inequalities for the cross-validation estimate of the generalization error for subagged estimators, both for classification and regressor. General loss functions and class of predictors with both finite and infinite VC-dimension are considered. We slightly generalize the formalism introduced by \cite{DUD03} to cover a large variety of cross-validation procedures including leave-one-out cross-validation, $k$-fold cross-validation, hold-out cross-validation (or split sample), and the leave-$\upsilon$-out cross-validation. \bigskip \noindent An interesting consequence is that the probability upper bound is bounded by the minimum of a Hoeffding-type bound and a Vapnik-type bounds, and thus is smaller than 1 even for small learning set. Finally, we give a simple rule on how to subbag the predictor. \bigskip
Cornec Matthieu
No associations
LandOfFree
Estimating Subagging by cross-validation does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Estimating Subagging by cross-validation, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Estimating Subagging by cross-validation will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-588878