Computing Optimal Attribute Weight Settings for Nearest Neighbor Algorithms
作者:Charles X. Ling, Hangdong Wang
摘要
Nearest neighbor (NN) learning algorithms, examples of the lazy learning paradigm, rely on a distance function to measure the similarity of testing examples with the stored training examples. Since certain attributes are more discriminative, while others can be less or totally irrelevant, attributes should be weighed differently in the distance function. Most previous studies on weight setting for NN learning algorithms are empirical. In this paper we describe our attempt on deciding theoretically optimal weights that minimize the predictive error for NN algorithms. Assuming a uniform distribution of examples in a 2-d continuous space, we first derive the average predictive error introduced by a linear classification boundary, and then determine the optimal weight setting for any polygonal classification region. Our theoretical results of optimal attribute weights can serve as a baseline or lower bound for comparing other empirical weight setting methods.
论文关键词:nearest neighbor learning algorithms, lazy learning, attribute weight setting, theoretical analyses
论文评审过程:
论文官网地址:https://doi.org/10.1023/A:1006560730186