Nearest neighbor classification is one of the simplest and most popular methods for statistical pattern recognition. A major issue in k-nearest neighbor classification is how to find an optimal value of the neighborhood parameter k. In practice, this value is generally estimated by the method of cross-validation. However, the ideal value of k in a classification problem not only depends on the entire data set, but also on the specific observation to be classified. Instead of using any single value of k, this paper studies results for a finite sequence of classifiers indexed by k. Along with the usual posterior probability estimates, a new measure, called the Bayesian measure of strength, is proposed and investigated in this paper as a measure of evidence for different classes. The results of these classifiers and their corresponding estimated misclassification probabilities are visually displayed using shaded strips. These plots provide an effective visualization of the evidence in favor of different classes when a given data point is to be classified. We also propose a simple weighted averaging technique that aggregates the results of different nearest neighbor classifiers to arrive at the final decision. Based on the analysis of several benchmark data sets, the proposed method is found to be better than using a single value of k.
引用
收藏
页码:1592 / 1602
页数:11
相关论文
共 40 条
[1]
Aho A.V., 1974, The Design and Analysis of Computer Algorithms
机构:
Indian Stat Inst, Theoret Stat & Math Unit, Kolkata 700035, W Bengal, IndiaIndian Stat Inst, Theoret Stat & Math Unit, Kolkata 700035, W Bengal, India
Chaudhuri, P
Marron, JS
论文数: 0引用数: 0
h-index: 0
机构:Indian Stat Inst, Theoret Stat & Math Unit, Kolkata 700035, W Bengal, India
机构:
Indian Stat Inst, Theoret Stat & Math Unit, Kolkata 700035, W Bengal, IndiaIndian Stat Inst, Theoret Stat & Math Unit, Kolkata 700035, W Bengal, India
Chaudhuri, P
Marron, JS
论文数: 0引用数: 0
h-index: 0
机构:Indian Stat Inst, Theoret Stat & Math Unit, Kolkata 700035, W Bengal, India