Local Classifier Weighting by Quadratic Programming

被引:39
作者
Cevikalp, Hakan [1 ]
Polikar, Robi [2 ]
机构
[1] Eskisehir Osmangazi Univ, Dept Elect & Elect Engn, TR-26480 Meselik, Eskisehir, Turkey
[2] Rowan Univ, Dept Elect & Comp Engn, Glassboro, NJ 08028 USA
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 10期
基金
美国国家科学基金会;
关键词
Classification; classifier fusion; combining multiple classifiers; ensemble learning; local classification accuracy; neural networks; quadratic programming;
D O I
10.1109/TNN.2008.2005301
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
It has been widely accepted that the classification accuracy can be improved by combining outputs of multiple classifiers. However, how to combine multiple classifiers with various (potentially conflicting) decisions is still an open problem. A rich collection of classifier combination procedures-many of which are heuristic in nature-have been developed for this goal. In this brief, we describe a dynamic approach to combine classifiers that have expertise in different regions of the input space. To this end, we use local classifier accuracy estimates to weight classifier outputs. Specifically, we estimate local recognition accuracies of classifiers near a query sample by utilizing its nearest neighbors, and then use these estimates to find the best weights of classifiers to label the query. The problem is formulated as a convex quadratic optimization problem, which returns optimal nonnegative classifier weights with respect to the chosen objective function, and the weights ensure that locally most accurate classifiers are weighted more heavily for labeling the query sample. Experimental results on several data sets indicate that the proposed weighting scheme outperforms other popular classifier combination schemes, particularly on problems with complex decision boundaries. Hence, the results indicate that local classification-accuracy-based combination techniques are well suited for decision making when the classifiers are trained by focusing on different regions of the input space.
引用
收藏
页码:1832 / +
页数:9
相关论文
共 24 条
[1]
Breiman L, 1996, MACH LEARN, V24, P49
[2]
A decision-theoretic generalization of on-line learning and an application to boosting [J].
Freund, Y ;
Schapire, RE .
JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1997, 55 (01) :119-139
[3]
Freund Y, 1996, ICML
[4]
Friedman J. H., 1977, ACM Transactions on Mathematical Software, V3, P209, DOI 10.1145/355744.355745
[5]
[6]
Multiple classifier fusion in probabilistic neural networks [J].
Grim, J ;
Kittler, J ;
Pudil, P ;
Somol, P .
PATTERN ANALYSIS AND APPLICATIONS, 2002, 5 (02) :221-233
[7]
HO TK, 1994, IEEE T PATTERN ANAL, V16, P66, DOI 10.1109/34.273716
[8]
A comparison of methods for multiclass support vector machines [J].
Hsu, CW ;
Lin, CJ .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2002, 13 (02) :415-425
[9]
INDYK P, 2004, HDB DISCRETE COMPUTA, pCH39
[10]
Adaptive Mixtures of Local Experts [J].
Jacobs, Robert A. ;
Jordan, Michael I. ;
Nowlan, Steven J. ;
Hinton, Geoffrey E. .
NEURAL COMPUTATION, 1991, 3 (01) :79-87