Bagging equalizes influence

被引:87
作者
Grandvalet, Y [1 ]
机构
[1] Univ Technol Compiegne, CNRS, UMR 6599, F-60206 Compiegne, France
关键词
bagging; influence; leverage; bias/variance;
D O I
10.1023/B:MACH.0000027783.34431.42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bagging constructs an estimator by averaging predictors trained on bootstrap samples. Bagged estimates almost consistently improve on the original predictor. It is thus important to understand the reasons for this success, and also for the occasional failures. It is widely believed that bagging is effective thanks to the variance reduction stemming from averaging predictors. However, seven years from its introduction, bagging is still not fully understood. This paper provides experimental evidence supporting the hypothesis that bagging stabilizes prediction by equalizing the influence of training examples. This effect is detailed in two different frameworks: estimation on the real line and regression. Bagging's improvements/ deteriorations are explained by the goodness/badness of highly influential examples, in situations where the usual variance reduction argument is at best questionable. Finally, reasons for the equalization effect are advanced. They support that other resampling strategies such as half-sampling should provide qualitatively identical effects while being computationally less demanding than bootstrap sampling.
引用
收藏
页码:251 / 270
页数:20
相关论文
共 26 条
[1]  
[Anonymous], EFFECT BAGGING VARIA
[2]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[3]  
Breiman L, 1996, ANN STAT, V24, P2350
[4]   Prediction games and arcing algorithms [J].
Breiman, L .
NEURAL COMPUTATION, 1999, 11 (07) :1493-1517
[5]  
Breiman L, 1998, ANN STAT, V26, P801
[6]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[7]  
BREIMAN L, 1996, 460 U CAL BERK STAT
[8]  
BUHLMANN P, 2000, SEM STAT ETH ZUR
[9]  
Burgess AN, 1997, ADV NEUR IN, V9, P382
[10]   An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization [J].
Dietterich, TG .
MACHINE LEARNING, 2000, 40 (02) :139-157