AN EXPERIMENTAL COMPARISON OF THE NEAREST-NEIGHBOR AND NEAREST-HYPERRECTANGLE ALGORITHMNS

被引:88
作者
WETTSCHERECK, D
DIETTERICH, TG
机构
关键词
EXEMPLAR-BASED LEARNING; INSTANCE-BASED LEARNING; NESTED GENERALIZED EXEMPLARS; NEAREST NEIGHBORS; FEATURE WEIGHTS;
D O I
10.1007/BF00994658
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Algorithms based on Nested Generalized Exemplar (NGE) theory (Salzberg, 1991) classify new data points by computing their distance to the nearest ''generalized exemplar'' (i.e., either a point or an axis-parallel rectangle). They combine the distance-based character of nearest neighbor (NN) classifiers with the axis-parallel rectangle representation employed in many rule-learning systems. An implementation of NGE was compared to the k-nearest neighbor (kNN) algorithm in 11 domains and found to be significantly inferior to kNN in 9 of them Several modifications of NGE were studied to understand the cause of its poor performance. These show that its performance can be substantially improved by preventing NGE from creating overlapping rectangles, while still allowing complete nesting of rectangles. Performance can be further improved by modifying the distance metric to allow weights on each of the features (Salzberg, 1991). Best results were obtained in this study when the weights were computed using mutual information between the features and the output class. The best version of NGE developed is a batch algorithm (BNGE FWMI) that has no user-tunable parameters. BNGE FWMI's performance is comparable to the first-nearest neighbor algorithm (also incorporating feature weights). However, the k-nearest neighbor algorithm is still significantly superior to BNGE FWMI in 7 of the 11 domains, and inferior to it in only 2. We conclude that, even with our improvements, the NGE approach is very sensitive to the shape of the decision boundaries in classification problems. in domains where the decision boundaries are axis-parallel, the NGE approach can produce excellent generalization with interpretable hypotheses. In all domains tested, NGE algorithms require much less memory to store generalized exemplars than is required by NN algorithms.
引用
收藏
页码:5 / 27
页数:23
相关论文
共 12 条
[1]  
AHA DW, 1990, STUDY INSTANCE BASED
[2]  
BAKIRI G, 1991, THESIS OREGON STATE
[3]  
Dasarthy B., 1991, NEAREST NEIGHBOR NN
[4]   INTERNATIONAL APPLICATION OF A NEW PROBABILITY ALGORITHM FOR THE DIAGNOSIS OF CORONARY-ARTERY DISEASE [J].
DETRANO, R ;
JANOSI, A ;
STEINBRUNN, W ;
PFISTERER, M ;
SCHMID, JJ ;
SANDHU, S ;
GUPPY, KH ;
LEE, S ;
FROELICHER, V .
AMERICAN JOURNAL OF CARDIOLOGY, 1989, 64 (05) :304-310
[5]  
Duda R. O., 1973, PATTERN CLASSIFICATI, V3
[6]   VERY SIMPLE CLASSIFICATION RULES PERFORM WELL ON MOST COMMONLY USED DATASETS [J].
HOLTE, RC .
MACHINE LEARNING, 1993, 11 (01) :63-91
[7]  
Murphy P. M., 1994, UCI REPOSITORY MACHI
[8]  
QUINLAN JR, 1992, C4 5 PROGRAMS EMPIRI
[9]  
SALZBERG S, 1991, MACH LEARN, V6, P277
[10]   FUZZY MIN MAX NEURAL NETWORKS .1. CLASSIFICATION [J].
SIMPSON, PK .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (05) :776-786