Genetic Algorithm-Neural Network (GANN): a study of neural network activation functions and depth of genetic algorithm search applied to feature selection

被引:92
作者
Tong, Dong Ling [1 ]
Mintram, Robert [2 ]
机构
[1] Nottingham Trent Univ, John van Geest Canc Res Ctr, Sch Sci & Technol, Nottingham NG11 8NS, England
[2] Newcastle Univ, Newcastle Upon Tyne NE1 7RU, Tyne & Wear, England
关键词
Genetic Algorithm (GA); Artificial Neural Network (ANN); Activation function; GA evaluation; GA population; Feature selection; CLASSIFICATION; PREDICTION; DNA;
D O I
10.1007/s13042-010-0004-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hybrid genetic algorithms (GA) and artificial neural networks (ANN) are not new in the machine learning culture. Such hybrid systems have been shown to be very successful in classification and prediction problems. However, little attention has been focused on this architecture as a feature selection method and the consequent significance of the ANN activation function and the number of GA evaluations on the feature selection performance. The activation function is one of the core components of the ANN architecture and influences the learning and generalization capability of the network. Meanwhile the GA searches for an optimal ANN classifier given a set of chromosomes selected from those available. The objective of the GA is to combine the search for optimum chromosome choices with that of finding an optimum classifier for each choice. The process operates as a form of co-evolution with the eventual objective of finding an optimum chromosome selection rather than an optimum classifier. The selection of an optimum chromosome set is referred to in this paper as feature selection. Quantitative comparisons of four of the most commonly used ANN activation functions against ten GA evaluation step counts and three population sizes are presented. These studies employ four data sets with high dimension and low significant datum instances. That is to say that each datum has a high attribute count and the unusual or abnormal data are sparse within the data set. Results suggest that the hyperbolic tangent (tanh) activation function outperforms other common activation functions by extracting a smaller, but more significant feature set. Furthermore, it was found that fitness evaluation sizes ranging from 20,000 to 40,000 within populations ranging from 200 to 300, deliver optimum feature selection capability. Again, optimum in this sense meaning a smaller but more significant feature set.
引用
收藏
页码:75 / 87
页数:13
相关论文
共 18 条
[1]   GANN: Genetic algorithm neural networks for the detection of conserved combinations of features in DNA [J].
Beiko, RG ;
Charlebois, RL .
BMC BIOINFORMATICS, 2005, 6 (1)
[2]  
Bevilacqua V., 2006, Engineering Letters, V13, P335
[3]  
Cartwright H, 2008, CHAPTER EVOLUTIONARY, P113
[4]  
Cho HS, 2003, NANOTECH 03, V1
[5]  
DEJONG KA, 1991, LECT NOTES COMPUT SC, V496, P38
[6]   Molecular classification of cancer: Class discovery and class prediction by gene expression monitoring [J].
Golub, TR ;
Slonim, DK ;
Tamayo, P ;
Huard, C ;
Gaasenbeek, M ;
Mesirov, JP ;
Coller, H ;
Loh, ML ;
Downing, JR ;
Caligiuri, MA ;
Bloomfield, CD ;
Lander, ES .
SCIENCE, 1999, 286 (5439) :531-537
[7]   Using a genetic algorithm and a perceptron for feature selection and supervised class learning in DNA microarray data [J].
Karzynski, M ;
Mateos, A ;
Herrero, J ;
Dopazo, J .
ARTIFICIAL INTELLIGENCE REVIEW, 2003, 20 (1-2) :39-51
[8]   Classification and diagnostic prediction of cancers using gene expression profiling and artificial neural networks [J].
Khan, J ;
Wei, JS ;
Ringnér, M ;
Saal, LH ;
Ladanyi, M ;
Westermann, F ;
Berthold, F ;
Schwab, M ;
Antonescu, CR ;
Peterson, C ;
Meltzer, PS .
NATURE MEDICINE, 2001, 7 (06) :673-679
[9]  
Lin TC, 2006, LECT NOTES ARTIF INT, V4099, P1037
[10]  
Mitchell TM, 1997, AI MAG, V18, P11