Metric Embedding for Nearest Neighbor Classification

Statistics – Machine Learning

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

9 pages, 1 table

Scientific paper

The distance metric plays an important role in nearest neighbor (NN) classification. Usually the Euclidean distance metric is assumed or a Mahalanobis distance metric is optimized to improve the NN performance. In this paper, we study the problem of embedding arbitrary metric spaces into a Euclidean space with the goal to improve the accuracy of the NN classifier. We propose a solution by appealing to the framework of regularization in a reproducing kernel Hilbert space and prove a representer-like theorem for NN classification. The embedding function is then determined by solving a semidefinite program which has an interesting connection to the soft-margin linear binary support vector machine classifier. Although the main focus of this paper is to present a general, theoretical framework for metric embedding in a NN setting, we demonstrate the performance of the proposed method on some benchmark datasets and show that it performs better than the Mahalanobis metric learning algorithm in terms of leave-one-out and generalization errors.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Metric Embedding for Nearest Neighbor Classification does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Metric Embedding for Nearest Neighbor Classification, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Metric Embedding for Nearest Neighbor Classification will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-704219

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.