Mathematics – Statistics Theory
Scientific paper
2007-09-05
Bernoulli 2007, Vol. 13, No. 3, 799-819
Mathematics
Statistics Theory
Published at http://dx.doi.org/10.3150/07-BEJ5102 in the Bernoulli (http://isi.cbs.nl/bernoulli/) by the International Statist
Scientific paper
10.3150/07-BEJ5102
We investigate statistical properties for a broad class of modern kernel-based regression (KBR) methods. These kernel methods were developed during the last decade and are inspired by convex risk minimization in infinite-dimensional Hilbert spaces. One leading example is support vector regression. We first describe the relationship between the loss function $L$ of the KBR method and the tail of the response variable. We then establish the $L$-risk consistency for KBR which gives the mathematical justification for the statement that these methods are able to ``learn''. Then we consider robustness properties of such kernel methods. In particular, our results allow us to choose the loss function and the kernel to obtain computationally tractable and consistent KBR methods that have bounded influence functions. Furthermore, bounds for the bias and for the sensitivity curve, which is a finite sample version of the influence function, are developed, and the relationship between KBR and classical $M$ estimators is discussed.
Christmann Andreas
Steinwart Ingo
No associations
LandOfFree
Consistency and robustness of kernel-based regression in convex risk minimization does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Consistency and robustness of kernel-based regression in convex risk minimization, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Consistency and robustness of kernel-based regression in convex risk minimization will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-470328