l1-PENALIZED QUANTILE REGRESSION IN HIGH-DIMENSIONAL SPARSE MODELS

被引:370
作者
Belloni, Alexandre [1 ]
Chernozhukov, Victor [2 ,3 ]
机构
[1] Duke Univ, Fuqua Sch Business, Durham, NC 27708 USA
[2] MIT, Dept Econ, Cambridge, MA 02142 USA
[3] MIT, Ctr Operat Res, Cambridge, MA 02142 USA
基金
美国国家科学基金会;
关键词
Median regression; quantile regression; sparse models; LASSO; AGGREGATION; ESTIMATORS; SELECTION; RECOVERY;
D O I
10.1214/10-AOS827
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider median regression and, more generally, a possibly infinite collection of quantile regressions in high-dimensional sparse models. In these models, the number of regressors p is very large, possibly larger than the sample size n, but only at most s regressors have a nonzero impact on each conditional quantile of the response variable, where s grows more slowly than n. Since ordinary quantile regression is not consistent in this case, we consider l(1)-penalized quantile regression (l(1)-QR), which penalizes the l(1)-norm of regression coefficients, as well as the post-penalized QR estimator (post-l(1)-QR), which applies ordinary QR to the model selected by l(1)-QR. First, we show that under general conditions l(1)-QR is consistent at the near-oracle rate. root s/n root log(p boolean OR n), uniformly in the compact set u subset of (0, 1) of quantile indices. In deriving this result, we propose a partly pivotal, data-driven choice of the penalty level and show that it satisfies the requirements for achieving this rate. Second, we show that under similar conditions post-l(1)-QR is consistent at the near-oracle rate root s/n root log(p boolean OR n), uniformly over u, even if the l(1)-QR-selected models miss some components of the true models, and the rate could be even closer to the oracle rate otherwise. Third, we characterize conditions under which l(1)-QR contains the true model as a submodel, and derive bounds on the dimension of the selected model, uniformly over u; we also provide conditions under which hard-thresholding selects the minimal true model, uniformly over u.
引用
收藏
页码:82 / 130
页数:49
相关论文
共 34 条
  • [1] [Anonymous], 1995, THEORIE ANAL PROBABI
  • [2] BELLONI A, 2008, CONDITIONAL QUANTILE, P35443
  • [3] Belloni A., 2009, L1 PENALIZED QUANTIL
  • [4] BELLONI A, 2010, L1 PENALIZED QUANT S, P35443, DOI DOI 10.1214/10-AOS827SUPP
  • [5] BELLONI A, 2009, POST L1 PENALIZED ES, P35443
  • [6] ON THE COMPUTATIONAL COMPLEXITY OF MCMC-BASED ESTIMATORS IN LARGE SAMPLES
    Belloni, Alexandre
    Chernozhukov, Victor
    [J]. ANNALS OF STATISTICS, 2009, 37 (04) : 2011 - 2055
  • [7] Bertsimas Dimitris, 1997, Introduction to linear optimization, V6
  • [8] SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR
    Bickel, Peter J.
    Ritov, Ya'acov
    Tsybakov, Alexandre B.
    [J]. ANNALS OF STATISTICS, 2009, 37 (04) : 1705 - 1732
  • [9] CHANGES IN THE UNITED-STATES WAGE STRUCTURE 1963-1987 - APPLICATION OF QUANTILE REGRESSION
    BUCHINSKY, M
    [J]. ECONOMETRICA, 1994, 62 (02) : 405 - 458
  • [10] Sparsity oracle inequalities for the Lasso
    Bunea, Florentina
    Tsybakov, Alexandre
    Wegkamp, Marten
    [J]. ELECTRONIC JOURNAL OF STATISTICS, 2007, 1 : 169 - 194