Branch-and-bound algorithms for computing the best-subset regression models

被引:55
作者
Gatu, C
Kontoghiorghes, EJ
机构
[1] Univ Neuchatel, Dept Comp Sci, CH-2007 Neuchatel, Switzerland
[2] Univ Cyprus, Dept Publ & Business Adm, Nicosia, Cyprus
[3] Univ London Birkbeck Coll, Sch Comp Sci & Informat Syst, London WC1E 7HX, England
关键词
least squares; QR decomposition; subset regression;
D O I
10.1198/106186006X100290
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
An efficient branch-and-bound algorithm for computing the best-subset regression models is proposed. The algorithm avoids the computation of the whole regression tree that generates all possible subset models. It is formally shown that if the branch-and-bound test holds, then the current subtree together with its right-hand side subtrees are cut. This reduces significantly the computational burden of the proposed algorithm when compared to an existing leaps-and-bounds method which generates two trees. Specifically, the proposed algorithm, which is based on orthogonal transformations, outperforms by O(n(3)) the leaps-and-bounds strategy. The criteria used in identifying the best subsets are based on monotone functions of the residual sum of squares (RSS) such as R-2, adjusted R-2, mean square error of prediction, and C-p. Strategies and heuristics that improve the computational performance of the proposed algorithm are investigated. A computationally efficient heuristic version of the branch-and-bound strategy which decides to cut subtrees using a tolerance parameter is proposed. The heuristic algorithm derives models close to the best ones. However, it is shown analytically that the relative error of the RSS, and consequently the corresponding statistic, of the computed subsets is smaller than the value of the tolerance parameter which lies between zero and one. Computational results and experiments on random and real data arc presented and analyzed.
引用
收藏
页码:139 / 156
页数:18
相关论文
共 38 条
[1]   MEAN SQUARE ERROR OF PREDICTION AS A CRITERION FOR SELECTING VARIABLES [J].
ALLEN, DM .
TECHNOMETRICS, 1971, 13 (03) :469-&
[2]  
[Anonymous], PARALLEL ALGORITHMS
[3]  
[Anonymous], OPTIMIZATION HEURIST
[4]   BETTER SUBSET REGRESSION USING THE NONNEGATIVE GARROTE [J].
BREIMAN, L .
TECHNOMETRICS, 1995, 37 (04) :373-384
[5]  
CLARKE MRB, 1981, APPL STATIST, V30, P198, DOI DOI 10.2307/2346398
[6]   A FAST MODEL SELECTION PROCEDURE FOR LARGE FAMILIES OF MODELS [J].
EDWARDS, D ;
HAVRANEK, T .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1987, 82 (397) :205-213
[7]   Variable selection via nonconcave penalized likelihood and its oracle properties [J].
Fan, JQ ;
Li, RZ .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2001, 96 (456) :1348-1360
[8]  
Friedman J., 2001, The elements of statistical learning, V1, DOI DOI 10.1007/978-0-387-21606-5
[9]   REGRESSIONS BY LEAPS AND BOUNDS [J].
FURNIVAL, GM ;
WILSON, RW .
TECHNOMETRICS, 1974, 16 (04) :499-511
[10]   Parallel algorithms for computing all possible subset regression models using the QR decomposition [J].
Gatu, C ;
Kontoghiorghes, EJ .
PARALLEL COMPUTING, 2003, 29 (04) :505-521