On the generalization error of fixed combinations of classifiers

被引:3
作者
Anthony, Martin [1 ]
机构
[1] Univ London London Sch Econ & Polit Sci, Dept Math, London WC2A 2AE, England
关键词
computational learning; complexity of learning; generalization error; large margins;
D O I
10.1016/j.jcss.2006.10.017
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We consider the generalization error of concept learning when using a fixed Boolean function of the outputs of a number of different classifiers. Here, we take into account the 'margins' of each of the constituent classifiers. A special case is that in which the constituent classifiers are linear threshold functions (or perceptrons) and the fixed Boolean function is the majority function. This corresponds to a 'committee of perceptrons,' an artificial neural network (or circuit) consisting of a single layer of perceptrons (or linear threshold units) in which the output of the network is defined to be the majority output of the perceptrons. Recent work of Auer et al. studied the computational properties of such networks (where they were called 'parallel perceptrons'), proposed an incremental learning algorithm for them, and demonstrated empirically that the learning rule is effective. As a corollary of the results presented here, generalization error bounds are derived for this special case that provide further motivation for the use of this learning rule. (c) 2006 Elsevier Inc. All rights reserved.
引用
收藏
页码:725 / 734
页数:10
相关论文
共 20 条
[1]   Function learning from interpolation [J].
Anthony, M ;
Bartlett, PL .
COMBINATORICS PROBABILITY & COMPUTING, 2000, 9 (03) :213-225
[2]  
Anthony M, 2004, J MACH LEARN RES, V5, P189
[3]  
ANTHONY M., 1997, Computational learning theory, V30
[4]  
Anthony M., 1999, Neural Network Learning: Theoretical Foundations, V9
[5]  
Auer P, 2002, LECT NOTES COMPUT SC, V2415, P123
[6]  
AUER P, UNPUB LEARNING RULE
[7]   The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536
[8]   Enlarging the margins in perceptron decision trees [J].
Bennett, KP ;
Cristianini, N ;
Shawe-Taylor, J ;
Wu, DH .
MACHINE LEARNING, 2000, 41 (03) :295-313
[9]   LEARNABILITY AND THE VAPNIK-CHERVONENKIS DIMENSION [J].
BLUMER, A ;
EHRENFEUCHT, A ;
HAUSSLER, D ;
WARMUTH, MK .
JOURNAL OF THE ACM, 1989, 36 (04) :929-965
[10]  
DUDLEY RM, 1999, CAMBRIDGE STUD ADV M, V63