ON AN ASYMPTOTICALLY OPTIMAL ADAPTIVE CLASSIFIER DESIGN CRITERION

被引:1
作者
LEE, WT
TENORIO, MF
机构
[1] School of Electrical Engineering, Purdue University, West Lafayette
关键词
CLASSIFIER COMPLEXITY; CLASSIFIER DESIGN CRITERION; CLASSIFIER METRICS; HIDDEN LAYER OPTIMAL SIZING; MINIMUM DESCRIPTION LENGTH; NEURAL NETWORKS; VAPNIK-CHERVONEKIS DIMENSION (VCDIM);
D O I
10.1109/34.204915
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the central problems in classifier design is the estimation of classification error. The difficulty in estimating the error probability of a classifier is due to the unknown sample distribution and small number of training samples available. In this correspondence, we will present a new approach for solving this problem. In our model, there are two types of classification error: empirical and generalization error. The first is the error observed over the training samples, and the second is the discrepancy between the error probability and empirical error. In general, the error probability of a classifier can be bounded from above by the sum of empirical and generalization errors. Since both terms depend on classifier complexity, we need a proper measure for classifier complexity. In this research, we adopted the Vapnik and Chervonenkis dimension (VCdim) as such a measure. Based on this complexity measure, we have developed an estimate for generalization error. An optimal classifier design criterion (the generalized minimum empirical error criterion (GMEE)) has been proposed. The GMEE criterion consists of two terms: the empirical and the estimate of generalization error. This criterion is useful for optimal classifier design since the classifier that minimizes the criterion is the one with the smallest error probability. Furthermore, we prove that the GMEE criterion is GAMMA optimal. This means that the criterion can select the best classifier from GAMMA, which is a collection of classifiers with finite VCdim. As an application, the criterion is used to design the optimal neural network classifier. A corollary to the GAMMA optimality of neural network-based classifiers has been proven. Thus, our approach provides a theoretic foundation for the connectionist approach to optimal classifier design. Experimental results are given to validate this approach, followed by discussions and suggestions for future research.
引用
收藏
页码:312 / 318
页数:7
相关论文
共 28 条
[1]  
BARRON AR, 1989, P DECISION CONTR
[2]  
BAUM EB, 1988, ADV NEURAL INFO PROC
[3]  
BLUMER A, 1987, UCSCCRL8720 UC SANTZ
[4]  
COHN D, 1990, TR910304 U WASH
[5]   SMOOTHING NOISY DATA WITH SPLINE FUNCTIONS [J].
WAHBA, G .
NUMERISCHE MATHEMATIK, 1975, 24 (05) :383-393
[6]  
CYBENKO G, 1989, CSRD856 U ILL REP
[8]  
Duda R. O., 1973, PATTERN CLASSIFICATI, V3
[9]   CENTRAL LIMIT-THEOREMS FOR EMPIRICAL MEASURES [J].
DUDLEY, RM .
ANNALS OF PROBABILITY, 1978, 6 (06) :899-929
[10]   CONSIDERATIONS OF SAMPLE AND FEATURE SIZE [J].
FOLEY, DH .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1972, 18 (05) :618-+