Can threshold networks be trained directly?

被引:191
作者
Huang, GB [1 ]
Zhu, QY [1 ]
Mao, KZ [1 ]
Siew, CK [1 ]
Saratchandran, P [1 ]
Sundararajan, N [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
关键词
extreme learning machine (ELM); gradient descent method; threshold neural networks;
D O I
10.1109/TCSII.2005.857540
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Neural networks with threshold activation functions are highly desirable because of the ease of hardware implementation, However, the popular gradient-based learning algorithms cannot be directly used to train these networks as the threshold functions are nondifferentiable. Methods available in the literature mainly focus on approximating the threshold activation functions by using sigmoid functions. In this paper, we show theoretically that the recently developed extreme learning machine (ELM) algorithm can be used to train the neural networks with threshold functions directly instead of approximating them with sigmoid functions. Experimental results based on real-world benchmark regression problems demonstrate that the generalization performance obtained by ELM is better than other algorithms used in threshold networks. Also, the ELM method does not need control variables (manually tuned parameters) and is much faster.
引用
收藏
页码:187 / 191
页数:5
相关论文
共 13 条
  • [1] [Anonymous], 2004, 2004 IEEE INT JOINT
  • [2] The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network
    Bartlett, PL
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) : 525 - 536
  • [3] USING RANDOM WEIGHTS TO TRAIN MULTILAYER NETWORKS OF HARD-LIMITING UNITS
    BARTLETT, PL
    DOWNS, T
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (02): : 202 - 210
  • [4] Baum E. B., 1988, Journal of Complexity, V4, P193, DOI 10.1016/0885-064X(88)90020-9
  • [5] AN ITERATIVE METHOD FOR TRAINING MULTILAYER NETWORKS WITH THRESHOLD FUNCTIONS
    CORWIN, EM
    LOGAR, AM
    OLDHAM, WJB
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (03): : 507 - 508
  • [6] GOODMAN R, 1994, P IEEE NEUR NETW SIG, P219
  • [7] REPRESENTING AND LEARNING BOOLEAN FUNCTIONS OF MULTIVALUED FEATURES
    HAMPSON, SE
    VOLPER, DJ
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1990, 20 (01): : 67 - 80
  • [8] Learning capability and storage capacity of two-hidden-layer feedforward networks
    Huang, GB
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2003, 14 (02): : 274 - 281
  • [9] Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions
    Huang, GB
    Babri, HA
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1998, 9 (01): : 224 - 229
  • [10] Plagianakos VP, 2001, P IEEE INT JOINT C N