Issues in stacked generalization

被引:454
作者
Ting, KM [1 ]
Witten, IH
机构
[1] Deakin Univ, Sch Comp & Math, Geelong, Vic 3217, Australia
[2] Univ Waikato, Dept Comp Sci, Hamilton, New Zealand
关键词
D O I
10.1613/jair.594
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stacked generalization is a general method of using a high-level model to combine lower-level models to achieve greater predictive accuracy. In this paper we address two crucial issues which have been considered to be a `black art' in classification tasks ever since the introduction of stacked generalization in 1992 by Wolpert: the type of generalizer that is suitable to derive the higher-level model, and the kind of attributes that should be used as its input. We find that best results are obtained when the higher-level model combines the confidence (and not just the predictions) of the lower-level ones. We demonstrate the effectiveness of stacked generalization for combining three different types of learning algorithms for classification tasks. We also compare the performance of stacked generalization with majority vote and published results of arcing and bagging.
引用
收藏
页码:271 / 289
页数:19
相关论文
共 36 条
[1]  
AHA DW, 1991, MACH LEARN, V6, P37, DOI 10.1007/BF00153759
[2]   Error reduction through learning multiple descriptions [J].
Ali, KM ;
Pazzani, MJ .
MACHINE LEARNING, 1996, 24 (03) :173-202
[3]  
[Anonymous], ARTIFICIAL NEURAL NE
[4]  
[Anonymous], P 14 INT C MACH LEAR
[5]  
[Anonymous], P 12 INT C MACH LEAR
[6]  
[Anonymous], 1998, UCI repository of machine learning databases.
[7]  
Breiman L, 1996, MACH LEARN, V24, P49
[8]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[9]  
Breiman L., 1996, Bias, variance, and arcing classifiers
[10]  
Cestnik B., 1990, P EUR C ART INT, P147