COMPUTING 2ND DERIVATIVES IN FEEDFORWARD NETWORKS - A REVIEW

被引:51
作者
BUNTINE, WL
WEIGEND, AS
机构
[1] UNIV COLORADO,DEPT COMP SCI,BOULDER,CO 80309
[2] UNIV COLORADO,INST COGNIT SCI,BOULDER,CO 80309
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1994年 / 5卷 / 03期
关键词
D O I
10.1109/72.286919
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The calculation of second derivatives is required by recent training and analysis techniques of connectionist networks, such as the elimination of superfluous weights, and the estimation of confidence intervals both for weights and network outputs. We here review and develop exact and approximate algorithms for calculating second derivatives. For networks with \w\ weights, simply writing the full matrix of second derivatives requires O(\w\2) operations. For networks of radial basis units or sigmoid units, exact calculation of the necessary intermediate terms requires of the order of 2h + 2 backward/forward-propagation passes where h is the number of hidden units in the network. We also review and compare three approximations (ignoring some components of the second derivative, numerical differentiation, and scoring). Our algorithms apply to arbitrary activation functions, networks, and error functions (for instance, with connections that skip layers, or radial basis functions, or cross-entropy error and Softmax units, etc.).
引用
收藏
页码:480 / 488
页数:9
相关论文
共 25 条
  • [1] [Anonymous], 1987, LEARNING INTERNAL RE
  • [2] [Anonymous], 1986, NUMERICAL RECIPES
  • [3] Barron A.R., 1988, 20TH INTERFACE 88 S, P192
  • [4] BAUM EB, 1987, NEURAL INFORMATION P, P52
  • [5] BECKER S, 1988, 1988 P CONN MOD SUMM, P29
  • [6] EXACT CALCULATION OF THE HESSIAN MATRIX FOR THE MULTILAYER PERCEPTRON
    BISHOP, C
    [J]. NEURAL COMPUTATION, 1992, 4 (04) : 494 - 501
  • [7] BRIDLE JS, 1989, NEUROCOMPUTING ALGOR, P223
  • [8] Cun Y.L., 1993, ADV NEURAL INFORM PR, P156
  • [9] ELJAROUDI A, 1990, INT JOINT C NEURAL N
  • [10] FAHLMAN SE, 1988, 1988 P CONN MOD SUMM, P38