中国科学技术大学学报 ›› 2018, Vol. 48 ›› Issue (4): 331-340.DOI: 10.3969/j.issn.0253-2778.2018.04.009

• 论著 • 上一篇    下一篇

基于数据划分的k-近邻分类加速算法机理分析

宋云胜,王杰,梁吉业,   

  1. 1.山西大学计算机与信息技术学院,山西太原 030006;2.计算智能与中文信息处理教育部重点实验室(山西大学),山西太原 030006
  • 收稿日期:2017-09-19 修回日期:2018-04-11 出版日期:2018-04-30 发布日期:2018-04-30
  • 通讯作者: 梁吉业
  • 作者简介:宋云胜, 男, 1984年生, 博士生, 研究方向: 机器学习.E-mail: sys_sd@126.com
  • 基金资助:
    国家自然科学基金重点项目(U1435212, 61432011), 山西省重点科技攻关项目(MQ2014-09)资助.

Mechanism analysis of the accelerator for k-nearest neighbor algorithm based on data partition

SONG Yunsheng,WANG Jie, LIANG Jiye,   

  1. 1. School of Computer & Information Technology, Shanxi University, Taiyuan 030006;
  • Received:2017-09-19 Revised:2018-04-11 Online:2018-04-30 Published:2018-04-30

摘要: k-近邻(kNN)分类算法因具有不对数据分布做任何假设、操作简单且泛化性能较强的特点,在人脸识别、文本分类、情感分析等领域被广泛使用.kNN分类算法不需要训练过程,其简单存储训练实例并根据测试实例与存储的训练实例进行相似度比较来预测分类.由于kNN分类算法需要计算测试实例与所有训练实例之间的相似度,故难以高效地处理大规模数据.为此提出将寻找近邻的过程转化为一个优化问题,并给出了原始优化问题与使用数据划分优化问题的最优解下目标函数差异的估计.通过对此估计的理论分析表明,聚类划分可以有效的减小此差异,进而保证基于聚类的k-近邻分类(DC-kNN)算法具有较强的泛化性能.在公开数据集的实验结果显示,DC-kNN分类算法在很大程度上为测试实例提供了与原始kNN分类算法相同的k个近邻进而获得较高的分类精度.

关键词: k-近邻, 数据划分, 局部信息, 实例子集, 聚类

Abstract: Due to its absence of hypotheses for the underlying distributions of data, simple execution and strong generation ability, k-nearest neighbor classification algorithm (kNN) is widely used in face recognition, text classification, emotional analysis and other fields. kNN does not need the training process, but it only stores the training instances until the unlabeled instance appears, and executes the predicted process. However, kNN needs to compute the similarity between the unlabeled instance and all the training instances, hence it is difficult to deal with large-scale data. To overcome this difficulty, #br##br# the process of computing the nearest neighbors is converted to a constrained optimization problem, and an estimation is given of difference on the value of the objective function under the optimal solution with and without data partition. The theoretical analysis of this estimation indicates that data partition using clustering can reduce this difference, and the k-nearest neighbor algorithm based on clustering can have a strong generation ability. Experiment results on public datasets show that the k-nearest neighbor algorithm based on clustering can largely obtain the same nearest neighbors of raw kNN, thus obtaining higher classification accuracy.

Key words: k-nearest neighbor, data partition, local information, instance subset, clustering

中图分类号: