Statistics – Machine Learning
Scientific paper
2009-06-24
Philosophical Transactions of the Royal Society, Series A, vol. 367, pp. 4295-4312, 2009
Statistics
Machine Learning
18 pages, 6 figures, submitted for publication
Scientific paper
10.1098/rsta.2009.0161
In recent years, the spectral analysis of appropriately defined kernel matrices has emerged as a principled way to extract the low-dimensional structure often prevalent in high-dimensional data. Here we provide an introduction to spectral methods for linear and nonlinear dimension reduction, emphasizing ways to overcome the computational limitations currently faced by practitioners with massive datasets. In particular, a data subsampling or landmark selection process is often employed to construct a kernel based on partial information, followed by an approximate spectral analysis termed the Nystrom extension. We provide a quantitative framework to analyse this procedure, and use it to demonstrate algorithmic performance bounds on a range of practical approaches designed to optimize the landmark selection process. We compare the practical implications of these bounds by way of real-world examples drawn from the field of computer vision, whereby low-dimensional manifold structure is shown to emerge from high-dimensional video data streams.
Belabbas Mohamed-Ali
Wolfe Patrick J.
No associations
LandOfFree
On landmark selection and sampling in high-dimensional data analysis does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with On landmark selection and sampling in high-dimensional data analysis, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and On landmark selection and sampling in high-dimensional data analysis will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-536643