Manifold Regularized Discriminative Nonnegative Matrix Factorization With Fast Gradient Descent

被引:274
作者
Guan, Naiyang [1 ]
Tao, Dacheng [2 ]
Luo, Zhigang [1 ]
Yuan, Bo [3 ]
机构
[1] Natl Univ Def Technol, Sch Comp Sci, Changsha 410073, Hunan, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Quantum Computat & Intelligent Syst, Sydney, NSW 2007, Australia
[3] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Gradient descent; nonnegative matrix factorization (NMF); manifold regularization; RECOGNITION; PARTS; FRAMEWORK; OBJECTS;
D O I
10.1109/TIP.2011.2105496
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nonnegative matrix factorization (NMF) has become a popular data-representation method and has been widely used in image processing and pattern-recognition problems. This is because the learned bases can be interpreted as a natural parts-based representation of data and this interpretation is consistent with the psychological intuition of combining parts to form a whole. For practical classification tasks, however, NMF ignores both the local geometry of data and the discriminative information of different classes. In addition, existing research results show that the learned basis is unnecessarily parts-based because there is neither explicit nor implicit constraint to ensure the representation parts-based. In this paper, we introduce the manifold regularization and the margin maximization to NMF and obtain the manifold regularized discriminative NMF (MD-NMF) to overcome the aforementioned problems. The multiplicative update rule (MUR) can be applied to optimizing MD-NMF, but it converges slowly. In this paper, we propose a fast gradient descent (FGD) to optimize MD-NMF. FGD contains a Newton method that searches the optimal step length, and thus, FGD converges much faster than MUR. In addition, FGD includes MUR as a special case and can be applied to optimizing NMF and its variants. For a problem with 165 samples in R-1600, FGD converges in 28 s, while MUR requires 282 s. We also apply FGD in a variant of MD-NMF and experimental results confirm its efficiency. Experimental results on several face image datasets suggest the effectiveness of MD-NMF.
引用
收藏
页码:2030 / 2048
页数:19
相关论文
共 51 条
[1]   ON SQUARE ROOTS OF M-MATRICES [J].
ALEFELD, G ;
SCHNEIDER, N .
LINEAR ALGEBRA AND ITS APPLICATIONS, 1982, 42 (FEB) :119-132
[2]  
ANDO RK, 2006, ADV NEURAL INF PROCE, V4, P25
[3]  
[Anonymous], 1997, REGIONAL C SERIES MA
[4]   Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection [J].
Belhumeur, PN ;
Hespanha, JP ;
Kriegman, DJ .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (07) :711-720
[5]  
Belkin M, 2002, ADV NEUR IN, V14, P585
[6]  
Belkin M., 2003, THESIS U CHICAGO CHI
[7]  
Belkin M, 2006, J MACH LEARN RES, V7, P2399
[8]  
Bermudez A. J., 1994, SAVMA Symposium 1994 Proceedings., P1
[9]   GOLDSTEIN-LEVITIN-POLYAK GRADIENT PROJECTION METHOD [J].
BERTSEKAS, DP .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1976, 21 (02) :174-183
[10]  
BIAN W, 2009, ADV NEURAL INF PROCE, V7, P1