Sparse regression ensembles in infinite and finite hypothesis spaces

被引:32
作者
Rätsch, G
Demiriz, A
Bennett, KP
机构
[1] GMD FIRST, D-12489 Berlin, Germany
[2] Rensselaer Polytech Inst, Dept Decis Sci & Engn Syst, Troy, NY 12180 USA
[3] Rensselaer Polytech Inst, Dept Math Sci, Troy, NY 12180 USA
基金
美国国家科学基金会;
关键词
ensemble learning; boosting; regression; sparseness; semi-infinite programming;
D O I
10.1023/A:1013907905629
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We examine methods for constructing regression ensembles based on a linear program (LP). The ensemble regression function consists of linear combinations of base hypotheses generated by some boosting-type base learning algorithm. Unlike the classification case, for regression the set of possible hypotheses producible by the base learning algorithm may be infinite. We explicitly tackle the issue of how to define and solve ensemble regression when the hypothesis space is infinite. Our approach is based on a semi-infinite linear program that has an infinite number of constraints and a finite number of variables. We show that the regression problem is well posed for infinite hypothesis spaces in both the primal and dual spaces. Most importantly, we prove there exists an optimal solution to the infinite hypothesis space problem consisting of a finite number of hypothesis. We propose two algorithms for solving the infinite and finite hypothesis problems. One uses a column generation simplex-type algorithm and the other adopts an exponential barrier approach. Furthermore, we give sufficient conditions for the base learning algorithm and the hypothesis set to be used for infinite regression ensembles. Computational results show that these methods are extremely promising.
引用
收藏
页码:189 / 218
页数:30
相关论文
共 57 条
[1]  
[Anonymous], 2000, ICML
[2]  
[Anonymous], LNCS, DOI DOI 10.1007/BFB0020283
[3]  
[Anonymous], 2018, TIME SERIES PREDICTI
[4]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[5]  
BERTONI A, 1997, LNCS, V5, P343
[6]  
BERTSEKAS DP, 1995, NONLINEAR PROGRAMMIN
[7]   Parsimonious least norm approximation [J].
Bradley, PS ;
Mangasarian, OL ;
Rosen, JB .
COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 1998, 11 (01) :5-21
[8]   Prediction games and arcing algorithms [J].
Breiman, L .
NEURAL COMPUTATION, 1999, 11 (07) :1493-1517
[9]  
BREIMAN L, 504 U CAL STAT DEP
[10]  
BRENEMAN C, 2000, QSAR CELLS S COMP CH