Improved binary particle swarm optimization using catfish effect for feature selection

被引:218
作者
Chuang, Li-Yeh [2 ]
Tsai, Sheng-Wei [1 ]
Yang, Cheng-Hong [1 ,3 ]
机构
[1] Natl Kaohsiung Univ Appl Sci, Dept Elect Engn, Kaohsiung 80778, Taiwan
[2] I Shou Univ, Inst Biotechnol & Chem Engn, Kaohsiung 80041, Taiwan
[3] Toko Univ, Dept Network Syst, Chiayi 61363, Taiwan
关键词
Feature selection; Catfish binary particle swarm optimization; K-nearest neighbor; Leave-one-out cross-validation; ALGORITHMS;
D O I
10.1016/j.eswa.2011.04.057
中图分类号
TP18 [人工智能理论];
学科分类号
140502 [人工智能];
摘要
The feature selection process constitutes a commonly encountered problem of global combinatorial optimization. This process reduces the number of features by removing irrelevant, noisy, and redundant data, thus resulting in acceptable classification accuracy. Feature selection is a preprocessing technique with great importance in the fields of data analysis and information retrieval processing, pattern classification, and data mining applications. This paper presents a novel optimization algorithm called catfish binary particle swarm optimization (CatfishBPSO), in which the so-called catfish effect is applied to improve the performance of binary particle swarm optimization (BPSO). This effect is the result of the introduction of new particles into the search space ("catfish particles"), which replace particles with the worst fitness by the initialized at extreme points of the search space when the fitness of the global best particle has not improved for a number of consecutive iterations. In this study, the K-nearest neighbor (K-NN) method with leave-one-out cross-validation (LOOCV) was used to evaluate the quality of the solutions. CatfishBPSO was applied and compared to 10 classification problems taken from the literature. Experimental results show that CatfishBPSO simplifies the feature selection process effectively, and either obtains higher classification accuracy or uses fewer features than other feature selection methods. (C) 2011 Elsevier Ltd. All rights reserved.
引用
收藏
页码:12699 / 12707
页数:9
相关论文
共 39 条
[1]
Feature selection for structure-activity correlation using binary particle swarms [J].
Agrafiotis, DK ;
Cedeño, W .
JOURNAL OF MEDICINAL CHEMISTRY, 2002, 45 (05) :1098-1107
[2]
[Anonymous], 2001, SWARM INTELL-US
[3]
[Anonymous], 1991, NEAREST NEIGHB NORMS
[4]
[Anonymous], 1997, MACHINE LEARNING, MCGRAW-HILL SCIENCE/ENGINEERING/MATH
[5]
USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[6]
BELLO R, 2007, INTELLIGENT SYSTEMS, V691, P696
[7]
FAST GENETIC SELECTION OF FEATURES FOR NEURAL NETWORK CLASSIFIERS [J].
BRILL, FZ ;
BROWN, DE ;
MARTIN, WN .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (02) :324-328
[8]
Chuang LY, 2008, 2008 IEEE SWARM INTELLIGENCE SYMPOSIUM, P9, DOI 10.1109/APCSAC.2008.4625441
[9]
Conover WJ, 1999, Practical nonparametric statistics
[10]
NEAREST NEIGHBOR PATTERN CLASSIFICATION [J].
COVER, TM ;
HART, PE .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1967, 13 (01) :21-+