Supervised nonlinear dimensionality reduction for visualization and classification

被引:326
作者
Geng, X [1 ]
Zhan, DC [1 ]
Zhou, ZH [1 ]
机构
[1] Nanjing Univ, Natl Lab Novel Software Technol, Nanjing 210093, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS | 2005年 / 35卷 / 06期
关键词
classification; dimensionality reduction; manifold learning; supervised learning; visualization;
D O I
10.1109/TSMCB.2005.850151
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
When performing visualization and classification, people often confront the problem of dimensionality reduction. Isomap is one of the most promising nonlinear dimensionality reduction techniques. However, when Isomap is applied to real-world data, it shows some limitations, such as being sensitive to noise. In this paper, an improved version of Isomap, namely S-Isomap, is proposed. S-Isomap utilizes class information to guide the procedure of nonlinear dimensionality reduction. Such a kind of procedure is called supervised nonlinear dimensionality reduction. In S-Isomap, the neighborhood graph of the input data is constructed according to a certain kind of dissimilarity between data points, which is specially designed to integrate the class information. The dissimilarity has several good properties which help to discover the true neighborhood of the data and, thus, makes S-Isomap a robust technique for both visualization and classification, especially for real-world problems. In the visualization experiments.. S-Isomap is compared with Isomap, LLE, and WeightedIso. The results show that S-Isomap performs the best. In the classification experiments, S-Isomap is used as a preprocess of classification and compared with Isomap, WeightedIso, as well as some other well-established classification methods, including the K-nearest neighbor classifier, BP neural network, J4.8 decision tree, and SVM. The results reveal that S-Isomap excels compared to Isomap and WeightedIso in classification, and it is highly competitive with those well-known classification methods.
引用
收藏
页码:1098 / 1107
页数:10
相关论文
共 18 条
[1]  
Blake C.L., 1998, UCI repository of machine learning databases
[2]  
Cherkassky V, 1997, IEEE Trans Neural Netw, V8, P1564, DOI 10.1109/TNN.1997.641482
[3]   INDEPENDENT COMPONENT ANALYSIS, A NEW CONCEPT [J].
COMON, P .
SIGNAL PROCESSING, 1994, 36 (03) :287-314
[4]  
Cox T. F., 1994, MULTIDIMENSIONAL SCA
[5]  
Jolliffe I. T., 1986, Principal Component Analysis, DOI [DOI 10.1016/0169-7439(87)80084-9, 10.1007/0-387-22440-8_13, DOI 10.1007/0-387-22440-8_13]
[6]   SIMILARITY METRIC LEARNING FOR A VARIABLE-KERNEL CLASSIFIER [J].
LOWE, DG .
NEURAL COMPUTATION, 1995, 7 (01) :72-85
[7]  
Mardia K., 1979, MULTIVARIATE ANAL
[8]  
Quinlan J. R., 1993, C4 5 PROGRAMS MACHIN
[9]   Nonlinear dimensionality reduction by locally linear embedding [J].
Roweis, ST ;
Saul, LK .
SCIENCE, 2000, 290 (5500) :2323-+
[10]   LEARNING REPRESENTATIONS BY BACK-PROPAGATING ERRORS [J].
RUMELHART, DE ;
HINTON, GE ;
WILLIAMS, RJ .
NATURE, 1986, 323 (6088) :533-536