ESTIMATION OF HIGH-DIMENSIONAL LOW-RANK MATRICES

被引:211
作者
Rohde, Angelika [1 ]
Tsybakov, Alexandre B. [2 ]
机构
[1] Univ Hamburg, Fachbereich Math, D-20146 Hamburg, Germany
[2] CREST, Stat Lab, F-92240 Malakoff, France
关键词
High-dimensional low-rank matrices; empirical process; sparse recovery; Schatten norm; penalized least-squares estimator; quasi-convex Schatten class embeddings; TRACE-NORM; AGGREGATION; CONSISTENCY; SELECTION; ENTROPY;
D O I
10.1214/10-AOS860
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Suppose that we observe entries or, more generally, linear combinations of entries of an unknown m x T-matrix A corrupted by noise. We are particularly interested in the high-dimensional setting where the number mT of unknown entries can be much larger than the sample size N. Motivated by several applications, we consider estimation of matrix A under the assumption that it has small rank. This can be viewed as dimension reduction or sparsity assumption. In order to shrink toward a low-rank representation, we investigate penalized least squares estimators with a Schatten-p quasi-norm penalty term, p <= 1. We study these estimators under two possible assumptions-a modified version of the restricted isometry condition and a uniform bound on the ratio "empirical norm induced by the sampling operator/Frobenius norm." The main results are stated as nonasymptotic upper bounds on the prediction risk and on the Schatten-q risk of the estimators, where q is an element of [p, 2]. The rates that we obtain for the prediction risk are of the form rmIN (for m = T), up to logarithmic factors, where r is the rank of A. The particular examples of multi-task learning and matrix completion are worked out in detail. The proofs are based on tools from the theory of empirical processes. As a by-product, we derive bounds for the kth entropy numbers of the quasi-convex Schatten class embeddings, S-p(M) -> S-2(M) p < 1, which are of independent interest.
引用
收藏
页码:887 / 930
页数:44
相关论文
共 51 条
  • [1] Abernethy J, 2009, J MACH LEARN RES, V10, P803
  • [2] HIGH-DIMENSIONAL ANALYSIS OF SEMIDEFINITE RELAXATIONS FOR SPARSE PRINCIPAL COMPONENTS
    Amini, Arash A.
    Wainwright, Martin J.
    [J]. ANNALS OF STATISTICS, 2009, 37 (5B) : 2877 - 2921
  • [3] [Anonymous], ADV NEURAL INFORM PR
  • [4] Argyriou A., 2008, Advances in Neural Information Processing Systems, P25
  • [5] ARGYRIOU A., 2010, J MACH LEAR IN PRESS
  • [6] Convex multi-task feature learning
    Argyriou, Andreas
    Evgeniou, Theodoros
    Pontil, Massimiliano
    [J]. MACHINE LEARNING, 2008, 73 (03) : 243 - 272
  • [7] Bach FR, 2008, J MACH LEARN RES, V9, P1019
  • [8] Regularized estimation of large covariance matrices
    Bickel, Peter J.
    Levina, Elizaveta
    [J]. ANNALS OF STATISTICS, 2008, 36 (01) : 199 - 227
  • [9] SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR
    Bickel, Peter J.
    Ritov, Ya'acov
    Tsybakov, Alexandre B.
    [J]. ANNALS OF STATISTICS, 2009, 37 (04) : 1705 - 1732
  • [10] BUNEA F., 2010, ARXIV10042995