Solving local minima problem with large number of hidden nodes on two-layered feed-forward artificial neural networks

被引:56
作者
Choi, Bumghi [1 ]
Lee, Ju-Hong [1 ]
Kim, Deok-Hwan [2 ]
机构
[1] Inha Univ, Dept Comp Sci & Informat Engn, Inchon, South Korea
[2] Inha Univ, Dept Elect Engn, Inchon, South Korea
关键词
Backpropagation; Local minima; Hidden nodes; Target values; Separate learning;
D O I
10.1016/j.neucom.2008.04.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The gradient descent algorithms like backpropagation (BP) or its variations on multi-layered feed-forward networks are widely used in many applications. However, the most serious problem associated with the BP is local minima problem. Especially, an exceeding number of hidden nodes make the corresponding network deepen the local minima problem. We propose an algorithm which shows stable performance on training despite of the large number of hidden nodes. This algorithm is called separate learning algorithm in which hidden-to-output and input-to-hidden separately trained. Simulations on some benchmark problems have been performed to demonstrate the validity of the proposed method. (C) 2008 Elsevier B.V. All rights reserved.
引用
收藏
页码:3640 / 3643
页数:4
相关论文
共 12 条
  • [1] [Anonymous], 1990, Report No
  • [2] FAST STOCHASTIC GLOBAL OPTIMIZATION
    BILBRO, GL
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1994, 24 (04): : 684 - 689
  • [3] CETIN BC, 1993, P IEEE INT C NEUR NE, V2, P836
  • [4] HONT R, 1995, HDB GLOBAL OPTIMIZAT
  • [5] Global optimization by multilevel coordinate search
    Huyer, W
    Neumaier, A
    [J]. JOURNAL OF GLOBAL OPTIMIZATION, 1999, 14 (04) : 331 - 355
  • [6] LIPSCHITZIAN OPTIMIZATION WITHOUT THE LIPSCHITZ CONSTANT
    JONES, DR
    PERTTUNEN, CD
    STUCKMAN, BE
    [J]. JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 1993, 79 (01) : 157 - 181
  • [7] JORDANOV IN, 2004, 2 INT IEEE C INT SYS, V1, P34
  • [8] Deterministic nonmonotone strategies for effective training of multilayer perceptrons
    Plagianakos, VP
    Magoulas, GD
    Vrahatis, MN
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2002, 13 (06): : 1268 - 1284
  • [9] TOM A, 1994, J GLOBAL OPTIM, V5, P267
  • [10] Simulated annealing and weight decay in adaptive learning: The SARPROP algorithm
    Treadgold, NK
    Gedeon, TD
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1998, 9 (04): : 662 - 668