Learning k for kNN Classification

Date

2017

Authors

Zhang, S.
Li, X.
Zong, M.
Zhu, X.
Cheng, D.

Editors

Advisors

Journal Title

Journal ISSN

Volume Title

Type:

Journal article

Citation

ACM Transactions on Intelligent Systems and Technology, 2017; 8(3 article no. 43):1-19

Statement of Responsibility

Conference Name

Abstract

The K Nearest Neighbor (kNN) method has widely been used in the applications of data mining andmachine learning due to its simple implementation and distinguished performance. However, setting all test data with the same κvalue in the previous kNN methods has been proven to make these methods impractical in real applications. This article proposes to learn a correlation matrix to reconstruct test data points by training data to assign different κ values to different test data points, referred to as the Correlation Matrix kNN (CM-kNN for short) classification. Specifically, the least-squares loss function is employed to minimize the reconstruction error to reconstruct each test data point by all training data points. Then, a graph Laplacian regularizer is advocated to preserve the local structure of the data in the reconstruction process. Moreover, an ℓ1-norm regularizer and an ℓ2,1-norm regularizer are applied to learn different κ values for different test data and to result in low sparsity to remove the redundant/noisy feature from the reconstruction process, respectively. Besides for classification tasks, the kNNmethods (including our proposed CM-kNN method) are further utilized to regression and missing data imputation.We conducted sets of experiments for illustrating the efficiency, and experimental results showed that the proposed method was more accurate and efficient than existing kNN methods in data-mining applications, such as classification, regression, and missing data imputation. Copyright is held by the owner/author(s).

School/Discipline

Dissertation Note

Provenance

Description

Link to a related website: http://dl.acm.org/ft_gateway.cfm?id=2990508&type=pdf, Open Access via Unpaywall

Access Status

Rights

Copyright 2017 The Authors

License

Grant ID

Call number

Persistent link to this record