Conditional Mutual Information-Based Feature Selection Analyzing for Synergy and Redundancy

被引:59
作者
Cheng, Hongrong [1 ]
Qin, Zhiguang [1 ]
Feng, Chaosheng [1 ]
Wang, Yong [1 ]
Li, Fagen [1 ]
机构
[1] Univ Elect Sci & Technol, Dept Comp Sci, Chengdu, Sichuan, Peoples R China
基金
国家高技术研究发展计划(863计划);
关键词
Classification; feature selection; conditional mutual information; redundancy; interaction;
D O I
10.4218/etrij.11.0110.0237
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Battiti's mutual information feature selector (MIFS) and its variant algorithms are used for many classification applications. Since they ignore feature synergy, MIFS and its variants may cause a big bias when features are combined to cooperate together. Besides, MIFS and its variants estimate feature redundancy regardless of the corresponding classification task. In this paper, we propose an automated greedy feature selection algorithm called conditional mutual information-based feature selection (CMIFS). Based on the link between interaction information and conditional mutual information, CMIFS takes account of both redundancy and synergy interactions of features and identifies discriminative features. In addition, CMIFS combines feature redundancy evaluation with classification tasks. It can decrease the probability of mistaking important features as redundant features in searching process. The experimental results show that CMIFS can achieve higher best-classification-accuracy than MIFS and its variants, with the same or less (nearly 50%) number of features.
引用
收藏
页码:210 / 218
页数:9
相关论文
共 19 条
[1]   On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems [J].
Amaldi, E ;
Kann, V .
THEORETICAL COMPUTER SCIENCE, 1998, 209 (1-2) :237-260
[2]  
[Anonymous], 1993, Proceedings of the 13th International Joint Conference on Artificial Intelligence
[3]  
[Anonymous], PSYCHOMETRIKA
[4]   USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[5]  
Cover T.M., 2006, ELEMENTS INFORM THEO, V2nd ed
[6]  
Dash M., 1997, INTELL DATA ANAL, V1, P131
[7]   Normalized Mutual Information Feature Selection [J].
Estevez, Pablo. A. ;
Tesmer, Michel ;
Perez, Claudio A. ;
Zurada, Jacek A. .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (02) :189-201
[8]  
Fano R.M., 1961, Transmission of Information: A Statistical Theory of Communications
[9]  
[HUANG JinJie 黄金杰], 2008, [自动化学报, Acta Automatica Sinica], V34, P383
[10]  
Jakulin A., 2004, QUANTIFYING VISUALIZ