Optimization method based extreme learning machine for classification

被引:710
作者
Huang, Guang-Bin [1 ]
Ding, Xiaojian [1 ,2 ]
Zhou, Hongming [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
[2] Xi An Jiao Tong Univ, Sch Elect & Informat Engn, Xian 710049, Shaanxi, Peoples R China
关键词
Extreme learning machine; Support vector machine; Support vector network; ELM kernel; ELM feature space; Equivalence between ELM and SVM; Maximal margin; Minimal norm of weights; Primal and dual ELM networks; NETWORKS; APPROXIMATION;
D O I
10.1016/j.neucom.2010.02.019
中图分类号
TP18 [人工智能理论];
学科分类号
140502 [人工智能];
摘要
Extreme learning machine (ELM) as an emergent technology has shown its good performance in regression applications as well as in large dataset (and/or multi-label) classification applications The ELM theory shows that the hidden nodes of the generalized single-hidden layer feedforward networks (SLFNs) which need not be neuron alike can be randomly generated and the universal approximation capability of such SLFNs can be guaranteed This paper further studies ELM for classification in the aspect of the standard optimization method and extends ELM to a specific type of generalized SLFNs support vector network. This paper shows that (1) under the ELM learning framework SVM s maximal margin property and the minimal norm of weights theory of feedforward neural networks are actually consistent (2) from the standard optimization method point of view ELM for classification and SVM are equivalent but ELM has less optimization constraints due to its special separability feature (3) as analyzed in theory and further verified by the simulation results ELM for classification tends to achieve better generalization performance than traditional SVM ELM for classification is less sensitive to user specified parameters and can be implemented easily (C) 2010 Elsevier B V All rights reserved
引用
收藏
页码:155 / 163
页数:9
相关论文
共 26 条
[1]
[Anonymous], 2010, P 18 EUR S ART NEUR
[2]
[Anonymous], 1961, PRINCIPLES NEURODYNA
[3]
[Anonymous], P IEEE COMP SOC C BI
[4]
The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536
[5]
Blake C. L., 1998, Uci repository of machine learning databases
[6]
Canu S., 2005, Perception Systmes et Information
[7]
SUPPORT-VECTOR NETWORKS [J].
CORTES, C ;
VAPNIK, V .
MACHINE LEARNING, 1995, 20 (03) :273-297
[8]
Error Minimized Extreme Learning Machine With Growth of Hidden Nodes and Incremental Learning [J].
Feng, Guorui ;
Huang, Guang-Bin ;
Lin, Qingping ;
Gay, Robert .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (08) :1352-1357
[9]
Fletcher R., 1981, Practical methods of optimization, volume 2, Constrained Optimization, V2
[10]
Fung G., 2001, KDD-2001. Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, P77, DOI 10.1145/502512.502527