Stability analysis of a three-term backpropagation algorithm

被引:78
作者
Zweiri, YH [1 ]
Seneviratne, LD [1 ]
Althoefer, K [1 ]
机构
[1] Kings Coll London, Dept Mech Engn, London WC2R 2LS, England
关键词
neural networks; backpropagation; stability; jury test;
D O I
10.1016/j.neunet.2005.04.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
Efficient teaming by the backpropagation (BP) algorithm is required for many practical applications. The BP algorithm calculates the weight changes of artificial neural networks, and a common approach is to use a two-term algorithm consisting of a teaming rate (LR) and a momentum factor (MF). The major drawbacks of the two-term BP teaming algorithm are the problems of local minima and slow convergence speeds, which limit the scope for real-time applications. Recently the addition of an extra term, called a proportional factor (PF), to the two-term BP algorithm was proposed. The third increases the speed of the BP algorithm. However, the PF term also reduces the convergence of the BP algorithm, and criteria for evaluating convergence are required to facilitate the application of the three terms BP algorithm. This paper analyzes the convergence of the new three-term backpropagation algorithm. If the teaming parameters of the three-term BP algorithm satisfy the conditions given in this paper, then it is guaranteed that the system is stable and will converge to a local minimum. It is proved that if at least one of the eigenvalues of matrix F (compose of the Hessian of the cost function and the system Jacobian of the error vector at each iteration) is negative, then the system becomes unstable. Also the paper shows that all the local minima of the three-term BP algorithm cost function are stable. The relationship between the teaming parameters are established in this paper such that the stability conditions are met. (c) 2005 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1341 / 1347
页数:7
相关论文
共 19 条
[1]
ALEKSANDER I, 1995, INTRO NEURA COMPUTIN
[2]
Dynamics of multilayer networks in the vicinity of temporary minima [J].
Ampazis, N ;
Perantonis, SJ ;
Taylor, JG .
NEURAL NETWORKS, 1999, 12 (01) :43-58
[3]
CARL GL, 1996, NEUROCOMPUTING, V10, P7
[4]
Training multilayer neural networks using fast global learning algorithm - least-squares and penalized optimization methods [J].
Cho, SY ;
Chow, TWS .
NEUROCOMPUTING, 1999, 25 (1-3) :115-131
[6]
Goldberg J. L., 1991, MATRIX THEORY APPL
[7]
Learning algorithms based on linearization [J].
Hahnloser, R .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1998, 9 (03) :363-380
[8]
Horn R. A., 1986, Matrix analysis
[9]
Accelerating neural network training using weight extrapolations [J].
Kamarthi, SV ;
Pittner, S .
NEURAL NETWORKS, 1999, 12 (09) :1285-1299
[10]
COMPLETE SOLUTION OF THE LOCAL MINIMA IN THE XOR PROBLEM [J].
LISBOA, PJG ;
PERANTONIS, SJ .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1991, 2 (01) :119-124