Dynamics of learning near singularities in radial basis function networks

被引:35
作者
Wei, Haikun [1 ,2 ,3 ]
Amaria, Shun-ichi [1 ]
机构
[1] RIKEN, Brain Sci Inst, Amari Res Unit, Wako, Saitama 3510198, Japan
[2] Southeast Univ, Sch Automat, Nanjing 210096, Peoples R China
[3] Kyushu Inst Technol, Dept Brain Sci & Engn, Kitakyushu, Fukuoka 8080196, Japan
基金
中国国家自然科学基金;
关键词
Neural networks; RBF networks; Singularity; Dynamics of learning; Plateau phenomenon; Gradient descent learning; Stability analysis;
D O I
10.1016/j.neunet.2008.06.017
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The radial basis function (RBF) networks are one of the most widely used models for function approximation in the regression problem. In the learning paradigm, the best approximation is recursively or iteratively searched for based on observed data (teacher signals). One encounters difficulties in such a process when two component basis functions become identical, or when the magnitude of one component becomes null. In this case, the number of the components reduces by one, and then the reduced component recovers as the learning process proceeds further, provided such a component is necessary for the best approximation. Strange behaviors, especially the plateau phenomena, have been observed in dynamics of learning when such reduction occurs. There exist singularities in the space of parameters, and the above reduction takes place at the singular regions. This paper focuses on a detailed analysis of the dynamical behaviors of learning near the overlap and elimination singularities in RBF networks, based on the averaged learning equation that is applicable to both on-line and batch mode learning. We analyze the stability on the overlap singularity by solving the eigenvalues of the Hessian explicitly. Based on the stability analysis, we plot the analytical dynamic vector fields near the singularity, which are then compared to those real trajectories obtained by a numeric method. We also confirm the existence of the plateaus in both batch and on-line learning by simulation. (C) 2008 Elsevier Ltd. All rights reserved.
引用
收藏
页码:989 / 1005
页数:17
相关论文
共 14 条
[1]   A THEORY OF ADAPTIVE PATTERN CLASSIFIERS [J].
AMARI, S .
IEEE TRANSACTIONS ON ELECTRONIC COMPUTERS, 1967, EC16 (03) :299-+
[2]   Singularities affect dynamics of learning in neuromanifolds [J].
Amari, Shun-ichi ;
Park, Hyeyoung ;
Ozeki, Tomoko .
NEURAL COMPUTATION, 2006, 18 (05) :1007-1065
[3]   LEARNING BY ONLINE GRADIENT DESCENT [J].
BIEHL, M ;
SCHWARZE, H .
JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1995, 28 (03) :643-656
[4]  
COUSSEAU F, 2007, IEEE T NEUR IN PRESS
[5]   Dynamics of on-line learning in radial basis function networks [J].
Freeman, JAS ;
Saad, D .
PHYSICAL REVIEW E, 1997, 56 (01) :907-918
[6]   Online learning in radial basis function networks [J].
Freeman, JAS ;
Saad, D .
NEURAL COMPUTATION, 1997, 9 (07) :1601-1622
[7]   Local minima and plateaus in hierarchical structures of multilayer perceptrons [J].
Fukumizu, K ;
Amari, S .
NEURAL NETWORKS, 2000, 13 (03) :317-327
[8]   A regularity condition of the information matrix of a multilayer perceptron network [J].
Fukumizu, K .
NEURAL NETWORKS, 1996, 9 (05) :871-879
[10]  
Horn R. A., 1986, Matrix analysis