On adaptive learning rate that guarantees convergence in feedforward networks

被引:100
作者
Behera, Laxmidhar [1 ]
Kumar, Swagat [1 ]
Patnaik, Awhan [1 ]
机构
[1] Indian Inst Technol, Dept Elect Engn, Kanpur 208016, Uttar Pradesh, India
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2006年 / 17卷 / 05期
关键词
adaptive learning rate; backpropagation (BP); extended Kalman filtering (EKF); feedforward networks; Lyapunov function; Lyapunov stability theory; system-identification;
D O I
10.1109/TNN.2006.878121
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper investigates new learning algorithms (LF I and LF H) based on Lyapunov function for the training of feedforward neural networks. It is observed that such algorithms have interesting parallel with the popular backpropagation (BP) algorithm where the fixed learning rate is replaced by an adaptive learning rate computed using convergence theorem based on Lyapunov stability theory. LF 11, a modified version of LF 1, has been introduced with an.aim to avoid local minima. This modification also helps in improving the convergence speed in some cases. Conditions for achieving global minimum for these kind of algorithms have been studied in detail. The performances of the proposed algorithms are compared with BP algorithm and extended Kalman filtering (EKF) on three bench-mark function approximation problems: XOR, 3-bit parity, and 8-3 encoder. The comparisons are made in terms of number of learning iterations and computational time required for convergence. It is found that the proposed algorithms (LF I and H) are much faster in convergence than other two algorithms to attain same accuracy. Finally, the comparison is made on a complex two-dimensional (2-D) Gabor function and effect of adaptive learning rate for faster convergence is verified. In a nutshell, the investigations made in this paper help us better understand the learning procedure of feedforward neural networks in terms of adaptive learning rate, convergence speed, and local minima.
引用
收藏
页码:1116 / 1125
页数:10
相关论文
共 24 条
[1]   On adaptive trajectory tracking of a robot manipulator using inversion of its neural emulator [J].
Behera, L ;
Gopal, M ;
Chaudhury, S .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1996, 7 (06) :1401-1414
[2]   A fast training algorithm for neural networks [J].
Bilski, J ;
Rutkowski, L .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-ANALOG AND DIGITAL SIGNAL PROCESSING, 1998, 45 (06) :749-753
[3]   CONJUGATE-GRADIENT ALGORITHM FOR EFFICIENT TRAINING OF ARTIFICIAL NEURAL NETWORKS [J].
CHARALAMBOUS, C .
IEE PROCEEDINGS-G CIRCUITS DEVICES AND SYSTEMS, 1992, 139 (03) :301-310
[4]  
Chen C.-T., 1999, Linear system theory and design, V3rd
[5]   TRAINING FEEDFORWARD NETWORKS WITH THE MARQUARDT ALGORITHM [J].
HAGAN, MT ;
MENHAJ, MB .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (06) :989-993
[6]   A REAL-TIME LEARNING ALGORITHM FOR A MULTILAYERED NEURAL NETWORK BASED ON THE EXTENDED KALMAN FILTER [J].
IIGUNI, Y ;
SAKAI, H ;
TOKUMARU, H .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1992, 40 (04) :959-966
[7]   INCREASED RATES OF CONVERGENCE THROUGH LEARNING RATE ADAPTATION [J].
JACOBS, RA .
NEURAL NETWORKS, 1988, 1 (04) :295-307
[8]  
Kristic M., 1995, Nonlinear and Adaptive Control Design
[9]   Neighborhood based Levenberg-Marquardt algorithm for neural network training [J].
Lera, G ;
Pinzolas, M .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2002, 13 (05) :1200-1203
[10]   A sigma-pi-sigma neural network (SPSNN) [J].
Li, CK .
NEURAL PROCESSING LETTERS, 2003, 17 (01) :1-19