Performance enhancement of extreme learning machine for multi-category sparse data classification problems

被引:127
作者
Suresh, S. [1 ]
Saraswathi, S. [2 ]
Sundararajan, N. [3 ]
机构
[1] Nanyang Technol Univ, Sch Comp Engn, Singapore, Singapore
[2] Iowa State Univ, Laurence H Baker Ctr Bioinforrnat & Biol Stat, Ames, IA USA
[3] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
关键词
Neural network; Extreme learning machine; K-fold validation; Genetic algorithm; Multi-category sparse classification; Micro-array gene expression data; CAPABILITY;
D O I
10.1016/j.engappai.2010.06.009
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a performance enhancement scheme for the recently developed extreme learning machine (ELM) for multi-category sparse data classification problems. ELM is a single hidden layer neural network with good generalization capabilities and extremely fast learning capacity. In ELM, the input weights are randomly chosen and the output weights are analytically calculated. The generalization performance of the ELM algorithm for sparse data classification problem depends critically on three free parameters. They are, the number of hidden neurons, the input weights and the bias values which need to be optimally chosen. Selection of these parameters for the best performance of ELM involves a complex optimization problem. In this paper, we present a new, real-coded genetic algorithm approach called 'RCGA-ELM' to select the optimal number of hidden neurons, input weights and bias values which results in better performance. Two new genetic operators called 'network based operator' and 'weight based operator' are proposed to find a compact network with higher generalization performance. We also present an alternate and less computationally intensive approach called 'sparse-ELM'. Sparse-ELM searches for the best parameters of ELM using K-fold validation. A multi-class human cancer classification problem using micro-array gene expression data (which is sparse), is used for evaluating the performance of the two schemes. Results indicate that the proposed RCGA-ELM and sparse-ELM significantly improve ELM performance for sparse multi-category classification problems. (C) 2010 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1149 / 1157
页数:9
相关论文
共 12 条
[1]  
ALPAYDIN E, 2004, ASSESSING COMP CLASS, P327
[2]  
[Anonymous], 1994, GENETIC ALGORITHM DA
[3]  
Breiman L, 1996, ANN STAT, V24, P2350
[4]  
Huang GB, 2004, IEEE IJCNN, P985
[5]   Learning capability and storage capacity of two-hidden-layer feedforward networks [J].
Huang, GB .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2003, 14 (02) :274-281
[6]   Extreme learning machine: Theory and applications [J].
Huang, Guang-Bin ;
Zhu, Qin-Yu ;
Siew, Chee-Kheong .
NEUROCOMPUTING, 2006, 70 (1-3) :489-501
[7]   Universal approximation using incremental constructive feedforward networks with random hidden nodes [J].
Huang, Guang-Bin ;
Chen, Lei ;
Siew, Chee-Kheong .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2006, 17 (04) :879-892
[8]   Real-time learning capability of neural networks [J].
Huang, Guang-Bin ;
Zhu, Qin-Yu ;
Siew, Chee-Kheong .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2006, 17 (04) :863-878
[9]   Multiclass cancer diagnosis using tumor gene expression signatures [J].
Ramaswamy, S ;
Tamayo, P ;
Rifkin, R ;
Mukherjee, S ;
Yeang, CH ;
Angelo, M ;
Ladd, C ;
Reich, M ;
Latulippe, E ;
Mesirov, JP ;
Poggio, T ;
Gerald, W ;
Loda, M ;
Lander, ES ;
Golub, TR .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2001, 98 (26) :15149-15154
[10]   On comparing classifiers: Pitfalls to avoid and a recommended approach [J].
Salzberg, SL .
DATA MINING AND KNOWLEDGE DISCOVERY, 1997, 1 (03) :317-328