Mathematics – Statistics Theory
Scientific paper
2008-02-06
Journal of Machine Learning Research 10 (2009) 245-279
Mathematics
Statistics Theory
Scientific paper
Penalization procedures often suffer from their dependence on multiplying factors, whose optimal values are either unknown or hard to estimate from the data. We propose a completely data-driven calibration algorithm for this parameter in the least-squares regression framework, without assuming a particular shape for the penalty. Our algorithm relies on the concept of minimal penalty, recently introduced by Birge and Massart (2007) in the context of penalized least squares for Gaussian homoscedastic regression. On the positive side, the minimal penalty can be evaluated from the data themselves, leading to a data-driven estimation of an optimal penalty which can be used in practice; on the negative side, their approach heavily relies on the homoscedastic Gaussian nature of their stochastic framework. The purpose of this paper is twofold: stating a more general heuristics for designing a data-driven penalty (the slope heuristics) and proving that it works for penalized least-squares regression with a random design, even for heteroscedastic non-Gaussian data. For technical reasons, some exact mathematical results will be proved only for regressogram bin-width selection. This is at least a first step towards further results, since the approach and the method that we use are indeed general.
Arlot Sylvain
Massart Pascal
No associations
LandOfFree
Data-driven calibration of penalties for least-squares regression does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Data-driven calibration of penalties for least-squares regression, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Data-driven calibration of penalties for least-squares regression will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-560328