Efficient backpropagation learning using optimal learning rate and momentum

被引:114
作者
Yu, XH
Chen, GA
机构
[1] Southeast University
基金
中国国家自然科学基金;
关键词
multilayer feedforward neural networks; backpropagation learning;
D O I
10.1016/S0893-6080(96)00102-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper considers efficient backpropagation learning using dynamically optimal learning rate (LR) and momentum factor (MF). A family of approaches exploiting the derivatives with respect to the LR and MF is presented, which does not need to explicitly compute the first two order derivatives in weight space, but rather makes use of the information gathered from the forward and backward procedures. The computational and storage burden for estimating the optimal LR and MF at most triple that of the standard backpropagation algorithm (BPA); however, the backpropagation learning procedure can be accelerated with remarkable savings in running time. Extensive computer simulations provided in this paper indicate that at least a magnitude of savings in running time can be achieved using the present family of approaches. (C) 1997 Elsevier Science Ltd.
引用
收藏
页码:517 / 527
页数:11
相关论文
共 21 条
[1]  
ALLRED LG, 1990, P INT JOINT C NEUR N, V1, P702
[2]   1ST-ORDER AND 2ND-ORDER METHODS FOR LEARNING - BETWEEN STEEPEST DESCENT AND NEWTON METHOD [J].
BATTITI, R .
NEURAL COMPUTATION, 1992, 4 (02) :141-166
[3]  
BECKER S, 1988, 1988 P CONN MOD SUMM, P29
[4]  
CHEN GA, 1994, THESIS SE U NANJING
[5]  
HAYKIN S, 1994, NEURAL NETWORKS COMP, P138
[6]   Progress in supervised neural networks [J].
Hush, Don R. ;
Horne, Bill G. .
IEEE SIGNAL PROCESSING MAGAZINE, 1993, 10 (01) :8-39
[7]   INCREASED RATES OF CONVERGENCE THROUGH LEARNING RATE ADAPTATION [J].
JACOBS, RA .
NEURAL NETWORKS, 1988, 1 (04) :295-307
[8]   AN ADAPTIVE LEAST-SQUARES ALGORITHM FOR THE EFFICIENT TRAINING OF ARTIFICIAL NEURAL NETWORKS [J].
KOLLIAS, S ;
ANASTASSIOU, D .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS, 1989, 36 (08) :1092-1101
[9]  
Lippmann R. P., 1988, Computer Architecture News, V16, P7, DOI [10.1109/MASSP.1987.1165576, 10.1145/44571.44572]
[10]   A SCALED CONJUGATE-GRADIENT ALGORITHM FOR FAST SUPERVISED LEARNING [J].
MOLLER, MF .
NEURAL NETWORKS, 1993, 6 (04) :525-533