Scale-independent quality criteria for dimensionality reduction

被引:51
作者
Lee, John A. [1 ]
Verleysen, Michel [2 ,3 ]
机构
[1] Mol Imaging & Expt Radiotherapy Dept, B-1200 Brussels, Belgium
[2] Catholic Univ Louvain, Machine Learning Grp, B-1348 Louvain, Belgium
[3] Univ Paris 01, SAMOS MATISSE, F-75634 Paris 13, France
关键词
Dimensionality reduction; Embedding; Manifold learning; Quality assessment; COMPONENT ANALYSIS; PRESERVATION; EIGENMAPS;
D O I
10.1016/j.patrec.2010.04.013
中图分类号
TP18 [人工智能理论];
学科分类号
140502 [人工智能];
摘要
Dimensionality reduction aims at representing high-dimensional data in low-dimensional spaces, in order to facilitate their visual interpretation. Many techniques exist, ranging from simple linear projections to more complex nonlinear transformations. The large variety of methods emphasizes the need of quality criteria that allow for fair comparisons between them. This paper extends previous work about rank-based quality criteria and proposes to circumvent their scale dependency. Most dimensionality reduction techniques indeed rely on a scale parameter that distinguish between local and global data properties. Such a scale dependency can be similarly found in usual quality criteria: they assess the embedding quality on a certain scale. Experiments with various dimensionality reduction techniques eventually show the strengths and weaknesses of the proposed scale-independent criteria. (C) 2010 Elsevier B.V. All rights reserved.
引用
收藏
页码:2248 / 2257
页数:10
相关论文
共 46 条
[1]
[Anonymous], THESIS HELSINKI U TE
[2]
[Anonymous], 1952, Psychometrika
[3]
[Anonymous], 2000, Graph approximations to geodesics on embedded manifolds
[4]
QUANTIFYING THE NEIGHBORHOOD PRESERVATION OF SELF-ORGANIZING FEATURE MAPS [J].
BAUER, HU ;
PAWELZIK, KR .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (04) :570-579
[5]
Laplacian eigenmaps for dimensionality reduction and data representation [J].
Belkin, M ;
Niyogi, P .
NEURAL COMPUTATION, 2003, 15 (06) :1373-1396
[6]
BENGIO Y, 2003, 1239 U MONTR DEP INF
[7]
BRAND M, 2003, P INT WORKSH ART INT
[8]
Chen L., 2006, THESIS U PENNSYLVIAN
[9]
Local Multidimensional Scaling for Nonlinear Dimension Reduction, Graph Drawing, and Proximity Analysis [J].
Chen, Lisha ;
Buja, Andreas .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2009, 104 (485) :209-219
[10]
Curvilinear component analysis: A self-organizing neural network for nonlinear mapping of data sets [J].
Demartines, P ;
Herault, J .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1997, 8 (01) :148-154