Statistics – Machine Learning
Scientific paper
2007-06-24
Statistics
Machine Learning
9 pages, 1 table
Scientific paper
The distance metric plays an important role in nearest neighbor (NN) classification. Usually the Euclidean distance metric is assumed or a Mahalanobis distance metric is optimized to improve the NN performance. In this paper, we study the problem of embedding arbitrary metric spaces into a Euclidean space with the goal to improve the accuracy of the NN classifier. We propose a solution by appealing to the framework of regularization in a reproducing kernel Hilbert space and prove a representer-like theorem for NN classification. The embedding function is then determined by solving a semidefinite program which has an interesting connection to the soft-margin linear binary support vector machine classifier. Although the main focus of this paper is to present a general, theoretical framework for metric embedding in a NN setting, we demonstrate the performance of the proposed method on some benchmark datasets and show that it performs better than the Mahalanobis metric learning algorithm in terms of leave-one-out and generalization errors.
Lanckriet Gert R. G.
Sriperumbudur Bharath K.
No associations
LandOfFree
Metric Embedding for Nearest Neighbor Classification does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Metric Embedding for Nearest Neighbor Classification, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Metric Embedding for Nearest Neighbor Classification will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-704219