Globally convergent algorithms with local learning rates

被引:17
作者
Magoulas, GD [1 ]
Plagianakos, VP
Vrahatis, MN
机构
[1] Brunel Univ, Dept Informat Syst & Comp, Uxbridge UB8 3PH, Middx, England
[2] Univ Patras, Dept Math, GR-26110 Patras, Greece
[3] Univ Patras, UP Artificial Intelligence Res Ctr, GR-26110 Patras, Greece
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2002年 / 13卷 / 03期
关键词
backpropagation (BP) networks; batch training; endoscopy; globally convergent algorithms; gradient descent; local learning rate adaptation; Quickprop (Qprop); Silva-Almeida (SA) method;
D O I
10.1109/TNN.2002.1000148
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a new generalized theoretical result is presented that underpins the development of globally convergent first-order batch training algorithms which employ local learning rates. This result allows us to equip algorithms of this class with a strategy for adapting the overall direction of search to a descent one. In this way, a decrease of the batch-error measure at each training iteration is ensured, and convergence of the sequence of weight iterates to a local minimizer of the batch error function Is obtained from remote initial weights. The effectiveness of the theoretical result is illustrated in three application examples by comparing two well-known training algorithms with local learning rates to their globally convergent modifications.
引用
收藏
页码:774 / 779
页数:6
相关论文
共 34 条
[1]  
[Anonymous], P INT JOINT C NEUR N
[2]  
[Anonymous], 1989, Complex Syst
[3]  
[Anonymous], ACTA NUMERICA 1992
[5]   1ST-ORDER AND 2ND-ORDER METHODS FOR LEARNING - BETWEEN STEEPEST DESCENT AND NEWTON METHOD [J].
BATTITI, R .
NEURAL COMPUTATION, 1992, 4 (02) :141-166
[6]  
Chan L.-W., 1987, Computer Speech and Language, V2, P205, DOI 10.1016/0885-2308(87)90009-X
[7]  
Dennis J.E., 1996, NUMERICAL METHODS UN
[8]  
Fahlman S.E., 1989, P 1988 CONNECTIONIST, P38
[9]   Solving the N-bit parity problem using neural networks [J].
Hohil, ME ;
Liu, DR ;
Smith, SH .
NEURAL NETWORKS, 1999, 12 (09) :1321-1323
[10]   INCREASED RATES OF CONVERGENCE THROUGH LEARNING RATE ADAPTATION [J].
JACOBS, RA .
NEURAL NETWORKS, 1988, 1 (04) :295-307