A SCALED CONJUGATE-GRADIENT ALGORITHM FOR FAST SUPERVISED LEARNING

被引:2778
作者
MOLLER, MF
机构
关键词
FEEDFORWARD NEURAL NETWORK; SUPERVISED LEARNING; OPTIMIZATION; CONJUGATE GRADIENT ALGORITHMS;
D O I
10.1016/S0893-6080(05)80056-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A supervised learning algorithm (Scaled Conjugate Gradient, SCG) is introduced. The performance of SCG is benchmarked against that of the standard back propagation algorithm (BP) (Rumelhart, Hinton, & Williams, 1986), the conjugate gradient algorithm with line search (CGL) (Johansson, Dowla, & Goodman, 1990) and the one-step Broyden-Fletcher-Goldfarb-Shanno memoriless quasi-Newton algorithm (BFGS) (Battiti, 1990). SCG is fully-automated, includes no critical user-dependent parameters, and avoids a time consuming line search, which CGL and BFGS use in each iteration in order to determine an appropriate step size. Experiments show that SCG is considerably faster than BP, CGL, and BFGS.
引用
收藏
页码:525 / 533
页数:9
相关论文
共 19 条
[11]  
JUDD JS, 1987, COINS8760 U AMH TECH
[12]  
LECUN Y, 1989, CONNECTIONISM IN PERSPECTIVE, P143
[13]  
MOLLER MF, 1991, UNPUB SUPERVISED LEA
[14]  
MOLLER MF, 1990, 6TH P INT M YOUNG CO, V464, P184
[15]   RESTART PROCEDURES FOR CONJUGATE GRADIENT METHOD [J].
POWELL, MJD .
MATHEMATICAL PROGRAMMING, 1977, 12 (02) :241-254
[16]  
Tesauro G., 1987, Complex Systems, V1, P367
[17]  
WATROUS RL, 1987, 1ST P IEEE INT C NEU, V2, P619
[18]  
WILLIAMS P, 1990, COMMUNICATION
[19]  
YOSHIDA T, 1991, IJCNN91 SEATTLE