Gradient-based kernel dimension reduction for supervised learning

Statistics – Machine Learning

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

21 pages

Scientific paper

This paper proposes a novel kernel approach to linear dimension reduction for supervised learning. The purpose of the dimension reduction is to find directions in the input space to explain the output as effectively as possible. The proposed method uses an estimator for the gradient of regression function, based on the covariance operators on reproducing kernel Hilbert spaces. In comparison with other existing methods, the proposed one has wide applicability without strong assumptions on the distributions or the type of variables, and uses computationally simple eigendecomposition. Experimental results show that the proposed method successfully finds the effective directions with efficient computation.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Gradient-based kernel dimension reduction for supervised learning does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Gradient-based kernel dimension reduction for supervised learning, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Gradient-based kernel dimension reduction for supervised learning will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-688056

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.