Confidence-based classifier design

被引:47
作者
Li, Mingkun [1 ]
Sethi, Ishwar K. [1 ]
机构
[1] Oakland Univ, Dept Comp Sci & Engn, Rochester, MI 48309 USA
关键词
pattern classification; confidence-based classification; error estimation; reject option; support vector machines; dynamic bin width allocation;
D O I
10.1016/j.patcog.2006.01.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a new classifier design methodology, confidence-based classifier design, is proposed to design classifiers with controlled confidence. This methodology is under the guidance of two optimal classification theories, a new classification theory for designing optimal classifiers with controlled error rates and the C.K. Chow's optimal classification theory for designing optimal classifiers with controlled conditional error. The new methodology also takes advantage of the current well-developed classifier's probability preserving and ordering properties. It calibrates the output scores of current classifiers to the conditional error or error rates. Thus, it can either classify input samples or reject them according to the output scores of classifiers. It can achieve some reasonable performance even though it is not an optimal solution. An example is presented to implement the new methodology using support vector machines (SVMs). The empirical cumulative density function method is used to estimate error rates from the output scores of a trained SVM. Furthermore, a new dynamic bin width allocation method is proposed to estimate sample conditional error and this method adapts to the underlying probabilities. The experimental results clearly demonstrate the efficacy of the suggested classifier design methodology. (c) 2006 Pattern Recognition Society. Published by Elsevier Ltd. All rights reserved.
引用
收藏
页码:1230 / 1240
页数:11
相关论文
共 26 条
[1]  
[Anonymous], 1999, The Nature Statist. Learn. Theory
[2]  
[Anonymous], 2002, P 8 INT C KNOWL DISC
[3]  
[Anonymous], ADV LARGE MARGIN CLA
[4]  
Bain L., 1991, Introduction to probability and mathematical statistics
[5]   Multiple-prototype classifier design [J].
Bezdek, JC ;
Reichherzer, TR ;
Lim, GS ;
Attikiouzel, Y .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 1998, 28 (01) :67-79
[6]  
Blake C.L., 1998, UCI repository of machine learning databases
[7]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[8]   Choosing multiple parameters for support vector machines [J].
Chapelle, O ;
Vapnik, V ;
Bousquet, O ;
Mukherjee, S .
MACHINE LEARNING, 2002, 46 (1-3) :131-159
[9]  
CHOW CK, 1970, IEEE T INFORM THEORY, V16, P41, DOI 10.1109/TIT.1970.1054406
[10]  
Chow CK., 1957, IRE T ELECT COMPUT E, VEC-6, P247, DOI DOI 10.1109/TEC.1957.5222035