BOOSTING AND OTHER ENSEMBLE METHODS

被引:209
作者
DRUCKER, H [1 ]
CORTES, C [1 ]
JACKEL, LD [1 ]
LECUN, Y [1 ]
VAPNIK, V [1 ]
机构
[1] AT&T BELL LABS, HOLMDEL, NJ 07733 USA
关键词
D O I
10.1162/neco.1994.6.6.1289
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We compare the performance of three types of neural network-based ensemble techniques to that of a single neural network. The ensemble algorithms are two versions of boosting and committees of neural networks' trained independently. For each of the four algorithms, we experimentally determine the test and training error curves in an optical character recognition (OCR) problem as both a function of training set size and computational cost using three architectures. We show that a single machine is best for small training set size while for large training set size some version of boosting is best. However, for a given computational cost, boosting is always best. Furthermore, we show a surprising result for the original boosting algorithm: namely, that as the training set size increases, the training error decreases until it asymptotes to the test error rate. This has potential implications in the search for better training algorithms.
引用
收藏
页码:1289 / 1301
页数:13
相关论文
共 22 条
  • [1] [Anonymous], 1990, ADV NEURAL INF PROCE
  • [2] Atlas L., 1990, ADV NEURAL INFORM PR, V2, P566
  • [3] BAIRD HS, 1993, 2ND P IAPR INT C DOC, P20
  • [4] BAIRD HS, 1993, 2ND P ANN S DOC AN I
  • [5] BAUM EB, 1991, ADV NEURAL INFORMATI, V3, P904
  • [6] CONSENSUS THEORETIC CLASSIFICATION METHODS
    BENEDIKTSSON, JA
    SWAIN, PH
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1992, 22 (04): : 688 - 704
  • [7] DRUCKER H, 1993, NEURAL INFORMATION P, V5, P42
  • [8] Drucker H., 1993, INT J PATTERN RECOGN, V7, P704
  • [9] FRUEND Y, 1990, 3RD P ANN WORKSH COM, P202
  • [10] NEURAL NETWORK ENSEMBLES
    HANSEN, LK
    SALAMON, P
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1990, 12 (10) : 993 - 1001