On-line training of recurrent neural networks with continuous topology adaptation

被引:38
作者
Obradovic, D
机构
[1] Siemens AG, Corporate Research and Development, 81739 Munich
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1996年 / 7卷 / 01期
关键词
Feedback - Kalman filtering - Learning algorithms - Nonlinear systems - State estimation - Topology;
D O I
10.1109/72.478408
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a novel on-line procedure for training dynamic neural networks with input-output recurrences whose topology is continuously adjusted to the complexity of the target system dynamics. The latter is accomplished by changing the number of the elements of the network hidden layer whenever the existing topology cannot capture the dynamics presented by the new data. The training mechanism developed in this work is based on the suitably altered extended Kalman filter (EKF) algorithm which is simultaneously used for the network parameter adjustment and for its state estimation. The network itself consists of a single hidden layer with a Gaussian radial basis functions (GRBF's) and of a linear output layer. The choice of the GRBF is induced by the requirements of the online learning. The latter implies the network architecture which permits only local influence of the new data point in order not to forget the previously learned dynamics. The continuous topology adaptation is implemented in our algorithm to avoid memory and computational problems of using a regular grid of GRBF's which covers the network input space. Furthermore, we show that the resulting parameter increase can be handled ''smoothly'' without interfering with the already acquired information. In the case when the target system dynamics are changing over time, we show that a suitable forgetting factor can he used to ''unlearn'' the no-longer-relevant dynamics. The quality of the presented recurrent network training algorithm is demonstrated on the identification of nonlinear dynamic systems.
引用
收藏
页码:222 / 228
页数:7
相关论文
共 14 条
[1]  
CHEN S, 1991, SIGNAL PROCESSING, V22
[2]  
Farrell J., 1993, INTRO INTELLIGENT AU
[3]  
Gelb A., 1974, Applied Optimal Estimation
[4]  
Hertz J., 1991, Introduction to the Theory of Neural Computation
[5]  
Jazwinski A.H., 2007, Stochastic Processes and Filtering Theory
[6]  
LIVSTONE MM, 1992 P ACC CHIC, V2, P555
[7]  
Ljung L., 1987, System Identification: Theory for the User
[8]  
PARKER DB, 1987, 1ST P IEEE INT C NEU, V2, P593
[9]  
PATTERSON JW, 1990, IEEE T NEURAL NETWOR, V1
[10]  
POUSKORIS GV, 1991, P INT JOINT C NEURAL, V1, P771