Geometric Mean for Subspace Selection

被引:483
作者
Tao, Dacheng [1 ]
Li, Xuelong [2 ]
Wu, Xindong [3 ]
Maybank, Stephen J. [2 ]
机构
[1] Nanyang Technol Univ, Sch Comp Engn, Singapore, Singapore
[2] Univ London, Sch Comp Sci & Informat Syst, London WC1E 7HX, England
[3] Univ Vermont, Dept Comp Sci, Burlington, VT 05405 USA
关键词
Arithmetic mean; Fisher's linear discriminant analysis (FLDA); geometric mean; Kullback-Leibler (KL) divergence; machine learning; subspace selection (or dimensionality reduction); visualization; LINEAR DISCRIMINANT-ANALYSIS; DIMENSIONALITY REDUCTION; FACE RECOGNITION; LDA; CLASSIFICATION; DECOMPOSITION; CRITERION; IMAGE;
D O I
10.1109/TPAMI.2008.70
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Subspace selection approaches are powerful tools in pattern classification and data visualization. One of the most important subspace approaches is the linear dimensionality reduction step in the Fisher's linear discriminant analysis (FLDA), which has been successfully employed in many fields such as biometrics, bioinformatics, and multimedia information management. However, the linear dimensionality reduction step in FLDA has a critical drawback: for a classification task with c classes, if the dimension of the projected subspace is strictly lower than c - 1, the projection to a subspace tends to merge those classes, which are close together in the original feature space. If separate classes are sampled from Gaussian distributions, all with identical covariance matrices, then the linear dimensionality reduction step in FLDA maximizes the mean value of the Kullback-Leibler (KL) divergences between different classes. Based on this viewpoint, the geometric mean for subspace selection is studied in this paper. Three criteria are analyzed: 1) maximization of the geometric mean of the KL divergences, 2) maximization of the geometric mean of the normalized KL divergences, and 3) the combination of 1 and 2. Preliminary experimental results based on synthetic data, UCI Machine Learning Repository, and handwriting digits show that the third criterion is a potential discriminative subspace selection method, which significantly reduces the class separation problem in comparing with the linear dimensionality reduction step in FLDA and its several representative extensions.
引用
收藏
页码:260 / 274
页数:15
相关论文
共 44 条
  • [1] [Anonymous], 1998, UCI REPOSITORY MACHI
  • [2] [Anonymous], 1991, ELEMENTS INFORM THEO
  • [3] [Anonymous], PATTERN RECOGNITION
  • [4] Regularized Gaussian discriminant analysis through eigenvalue decomposition
    Bensmail, H
    Celeux, G
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1996, 91 (436) : 1743 - 1748
  • [5] BOUVEYRON C, 2007, COMM STAT THEORY MET, V36
  • [6] Boyd S., 2004, CONVEX OPTIMIZATION, DOI DOI 10.1017/CBO9780511804441
  • [7] Campbell N.A., 1984, Australian J. of Statistics, V26, P86, DOI DOI 10.1111/J.1467-842X.1984.TB01271.X
  • [8] Supervised tensor learning
    Dacheng Tao
    Xuelong Li
    Xindong Wu
    Weiming Hu
    Stephen J. Maybank
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2007, 13 (01) : 1 - 42
  • [9] DANIEL LS, 1999, IEEE T PATTERN ANAL, V21, P386
  • [10] Decell H. P. Jr., 1977, Computers & Mathematics with Applications, V3, P71, DOI 10.1016/0898-1221(77)90116-X