Sparse Extreme Learning Machine for Classification

被引:196
作者
Bai, Zuo [1 ]
Huang, Guang-Bin [1 ]
Wang, Danwei [1 ]
Wang, Han [1 ]
Westover, M. Brandon [2 ,3 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
[2] Massachusetts Gen Hosp, Boston, MA 02114 USA
[3] Harvard Univ, Sch Med, Boston, MA 02114 USA
关键词
Classification; extreme learning machine (ELM); quadratic programming (QP); sequential minimal optimization (SMO); sparse ELM; support vector machine (SVM); unified ELM; REGRESSION; ALGORITHM; NETWORKS;
D O I
10.1109/TCYB.2014.2298235
中图分类号
TP [自动化技术、计算机技术];
学科分类号
080201 [机械制造及其自动化];
摘要
Extreme learning machine (ELM) was initially proposed for single-hidden-layer feedforward neural networks (SLFNs). In the hidden layer (feature mapping), nodes are randomly generated independently of training data. Furthermore, a unified ELM was proposed, providing a single framework to simplify and unify different learning methods, such as SLFNs, least square support vector machines, proximal support vector machines, and so on. However, the solution of unified ELM is dense, and thus, usually plenty of storage space and testing time are required for large-scale applications. In this paper, a sparse ELM is proposed as an alternative solution for classification, reducing storage space and testing time. In addition, unified ELM obtains the solution by matrix inversion, whose computational complexity is between quadratic and cubic with respect to the training size. It still requires plenty of training time for large-scale problems, even though it is much faster than many other traditional methods. In this paper, an efficient training algorithm is specifically developed for sparse ELM. The quadratic programming problem involved in sparse ELM is divided into a series of smallest possible sub-problems, each of which are solved analytically. Compared with SVM, sparse ELM obtains better generalization performance with much faster training speed. Compared with unified ELM, sparse ELM achieves similar generalization performance for binary classification applications, and when dealing with large-scale binary classification problems, sparse ELM realizes even faster training speed than unified ELM.
引用
收藏
页码:1858 / 1870
页数:13
相关论文
共 33 条
[1]
Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays [J].
Alon, U ;
Barkai, N ;
Notterman, DA ;
Gish, K ;
Ybarra, S ;
Mack, D ;
Levine, AJ .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 1999, 96 (12) :6745-6750
[2]
[Anonymous], 2010, P 18 EUR S ART NEUR
[3]
[Anonymous], 1999, The Nature Statist. Learn. Theory
[4]
[Anonymous], 2003, PRACTICAL GUIDE SUPP
[5]
[Anonymous], 1999, SPRINGER SCI
[6]
Canu S., 2005, SVM KERNEL METHODS M
[7]
SUPPORT-VECTOR NETWORKS [J].
CORTES, C ;
VAPNIK, V .
MACHINE LEARNING, 1995, 20 (03) :273-297
[8]
Drucker H, 1997, ADV NEUR IN, V9, P155
[9]
Fletcher R., 1981, Practical methods of optimization, volume 2, Constrained Optimization, V2
[10]
Frank A., 2010, UCI machine learning repository, V213