Efficient classifiers for multi-class classification problems

被引:22
作者
Lin, Hung-Yi [1 ]
机构
[1] Natl Taichung Univ Sci & Technol, Dept Distribut Management, Taichung, Taiwan
关键词
Multivariate analysis; Multi-class problems; Feature evaluation; Feature selection; Feature extraction; Inductive learning; FEATURE-SELECTION; INFORMATION; DISCOVERY; CRITERIA;
D O I
10.1016/j.dss.2012.02.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Classification problems have become more complex and intricate in modern applications in the face of continuous data explosion. In addition to great quantities of features and large numbers of instances, modern classification applications are continuously developed with multiple classes (objectives). The ever-increasing growth in data quantity and computation complexity has largely deteriorated the performance and accuracy of classification models. In order to deal with such situations, multivariate statistical analyses are adopted in this paper. Multivariate statistical analyses have two advantages. First, they can explore the relationships between variables and find the most characterizing features of the observed data. Second, they can solve problems which are stalled by high dimensionality. In this paper, the first advantage is applied to the selection of relevant features and the second is employed to generate the multivariate classifier. Experimental results show that our model can significantly improve classification training time by combining a compact subset of relevant features without the loss of accuracy in multi-class classification problems. In addition, the discrimination degree of our classifier outperforms other conventional classifiers. (C) 2012 Elsevier B.V. All rights reserved.
引用
收藏
页码:473 / 481
页数:9
相关论文
共 45 条
[31]  
Orsenigo C., 2003, IMA Journal of Management Mathematics, V14, P221, DOI 10.1093/imaman/14.3.221
[32]   Identification and classification of conserved RNA secondary structures in the human genome [J].
Pedersen, Jakob Skou ;
Bejerano, Gill ;
Siepel, Adam ;
Rosenbloom, Kate ;
Lindblad-Toh, Kerstin ;
Lander, Eric S. ;
Kent, Jim ;
Miller, Webb ;
Haussler, David .
PLOS COMPUTATIONAL BIOLOGY, 2006, 2 (04) :251-262
[33]   Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy [J].
Peng, HC ;
Long, FH ;
Ding, C .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2005, 27 (08) :1226-1238
[34]   Anomaly Detection: A Survey [J].
Chandola, Varun ;
Banerjee, Arindam ;
Kumar, Vipin .
ACM COMPUTING SURVEYS, 2009, 41 (03)
[35]   Web Page Classification: Features and Algorithms [J].
Qi, Xiaoguang ;
Davison, Brian D. .
ACM COMPUTING SURVEYS, 2009, 41 (02)
[36]   Theoretical comparison between the Gini Index and Information Gain criteria [J].
Raileanu, LE ;
Stoffel, K .
ANNALS OF MATHEMATICS AND ARTIFICIAL INTELLIGENCE, 2004, 41 (01) :77-93
[37]  
Siepel A., 2004, P 8 ANN INT C COMP B
[38]   Combining multiple feature selection methods for stock prediction: Union, intersection, and multi-intersection approaches [J].
Tsai, Chih-Fong ;
Hsiao, Yu-Chieh .
DECISION SUPPORT SYSTEMS, 2010, 50 (01) :258-269
[39]   Feature selection in bankruptcy prediction [J].
Tsai, Chih-Fong .
KNOWLEDGE-BASED SYSTEMS, 2009, 22 (02) :120-127
[40]   Evolutionary-based feature selection approaches with new criteria for data mining: A case study of credit approval data [J].
Wang, Chia-Ming ;
Huang, Yin-Fu .
EXPERT SYSTEMS WITH APPLICATIONS, 2009, 36 (03) :5900-5908