High-Dimensional Feature Selection by Feature-Wise Non-Linear Lasso

Statistics – Machine Learning

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

13 pages

Scientific paper

The goal of supervised feature selection is to find a subset of input features that are responsible for predicting output values. The least absolute shrinkage and selection operator (Lasso) allows computationally efficient feature selection based on linear dependency between input features and output values. In this paper, we consider a feature-wise kernelized Lasso for capturing non-linear input-output dependency. We first show that, with particular choices of kernel functions,non-redundant features with strong statistical dependence on output values can be found in terms of kernel-based independence measures. We then show that the globally optimal solution can be efficiently computed; this makes the approach scalable to high-dimensional problems. The effectiveness of the proposed method is demonstrated through feature selection experiments with thousands of features.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

High-Dimensional Feature Selection by Feature-Wise Non-Linear Lasso does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with High-Dimensional Feature Selection by Feature-Wise Non-Linear Lasso, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and High-Dimensional Feature Selection by Feature-Wise Non-Linear Lasso will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-188536

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.