Evaluating feature selection methods for learning in data mining applications

被引:74
作者
Piramuthu, S [1 ]
机构
[1] Univ Florida, Gainesville, FL 32611 USA
关键词
artificial intelligence; feature selection; decision trees; credit risk analysis;
D O I
10.1016/s0377-2217(02)00911-6
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
Recent advances in computing technology in terms of speed, cost, as well as access to tremendous amounts of computing power and the ability to process huge amounts of data in reasonable time has spurred increased interest in data mining applications to extract useful knowledge from data. Machine learning has been one of the methods used in most of these data mining applications. It is widely acknowledged that about 80% of the resources in a majority of data mining applications are spent on cleaning and preprocessing the data. However, there have been relatively few studies on preprocessing data used as input in these data mining systems. In this study, we evaluate several inter-class as well as probabilistic distance-based feature selection methods as to their effectiveness in preprocessing input data for inducing decision trees. We use real-world data to evaluate these feature selection methods. Results from this study show that inter-class distance measures result in better performance compared to probabilistic measures, in general. (C) 2003 Elsevier B.V. All rights reserved.
引用
收藏
页码:483 / 494
页数:12
相关论文
共 39 条
[1]   INFORMATION CHOICE AND UTILIZATION IN AN EXPERIMENT ON DEFAULT PREDICTION [J].
ABDELKHALIK, AR ;
ELSHESHAI, KM .
JOURNAL OF ACCOUNTING RESEARCH, 1980, 18 (02) :325-342
[2]  
ALMUALLIM H, 1991, PROCEEDINGS : NINTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOLS 1 AND 2, P547
[3]  
[Anonymous], 1995, THESIS STANFORD U
[4]   USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[5]  
Bradley P. S., 1998, INFORMS Journal on Computing, V10, P209, DOI 10.1287/ijoc.10.2.209
[6]   BEST 2 INDEPENDENT MEASUREMENTS ARE NOT 2 BEST [J].
COVER, TM .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1974, SMC4 (01) :116-117
[7]  
Devijver P., 1982, PATTERN RECOGN
[8]   ON CHOICE OF VARIABLES IN CLASSIFICATION PROBLEMS WITH DICHOTOMOUS VARIABLES [J].
ELASHOFF, JD ;
ELASHOFF, RM ;
GOLDMAN, GE .
BIOMETRIKA, 1967, 54 :668-&
[9]  
EOLMAA T, 1994, P EUR C MACH LEARN, P351
[10]  
Fukunaga K., 1972, Introduction to statistical pattern recognition