A COMPARATIVE STUDY OF DATA SAMPLING TECHNIQUES FOR CONSTRUCTING NEURAL NETWORK ENSEMBLES

被引:19
作者
Akhand, M. A. H. [1 ]
Islam, M. D. Monirul [1 ]
Murase, Kazuyuki [1 ,2 ]
机构
[1] Univ Fukui, Grad Sch Engn, Fukui 9108507, Japan
[2] Univ Fukui, Res & Educ Program Life Sci, Fukui 9108507, Japan
关键词
Neural network ensemble; generalization; diversity; bagging; boosting; negative correlation learning; random subspace method; GRADIENT LEARNING ALGORITHM; CLASSIFICATION;
D O I
10.1142/S0129065709001859
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ensembles with several classifiers (such as neural networks or decision trees) are widely used to improve the generalization performance over a single classifier. Proper diversity among component classifiers is considered an important parameter for ensemble construction so that failure of one may be compensated by others. Among various approaches, data sampling, i.e., different data sets for different classifiers, is found more effective than other approaches. A number of ensemble methods have been proposed under the umbrella of data sampling in which some are constrained to neural networks or decision trees and others are commonly applicable to both types of classifiers. We studied prominent data sampling techniques for neural network ensembles, and then experimentally evaluated their effectiveness on a common test ground. Based on overlap and uncover, the relation between generalization and diversity is presented. Eight ensemble methods were tested on 30 benchmark classification problems. We found that bagging and boosting, the pioneer ensemble methods, are still better than most of the other proposed methods. However, negative correlation learning that implicitly encourages different networks to different training spaces is shown as better or at least comparable to bagging and boosting that explicitly create different training spaces.
引用
收藏
页码:67 / 89
页数:23
相关论文
共 30 条
[1]   AN ADAPTIVE CONJUGATE-GRADIENT LEARNING ALGORITHM FOR EFFICIENT TRAINING OF NEURAL NETWORKS [J].
ADELI, H ;
HUNG, SL .
APPLIED MATHEMATICS AND COMPUTATION, 1994, 62 (01) :81-102
[2]   A CONCURRENT ADAPTIVE CONJUGATE-GRADIENT LEARNING ALGORITHM ON MIMD SHARED-MEMORY MACHINES [J].
ADELI, H ;
HUNG, SL .
INTERNATIONAL JOURNAL OF SUPERCOMPUTER APPLICATIONS AND HIGH PERFORMANCE COMPUTING, 1993, 7 (02) :155-165
[3]   A comparison of decision tree ensemble creation techniques [J].
Banfield, Robert E. ;
Hall, Lawrence O. ;
Bowyer, Kevin W. ;
Kegelmeyer, W. P. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2007, 29 (01) :173-180
[4]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[5]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[6]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[7]   Randomizing outputs to increase prediction accuracy [J].
Breiman, L .
MACHINE LEARNING, 2000, 40 (03) :229-242
[8]  
Brown G., 2005, Information Fusion, V6, P99
[9]  
Chandra A, 2006, STUD COMP INTELL, V16, P429
[10]  
DePasquale J, 2007, LECT NOTES COMPUT SC, V4472, P251