RBF nets, mixture experts, and Bayesian Ying-Yang learning

被引:76
作者
Xu, L [1 ]
机构
[1] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, NT, Peoples R China
关键词
radial basis function network; mixture of experts; Bayesian Ying-Yang learning; model selection; coordinated competitive learning; fast computation; adaptive algorithm; financial time series; curve fitting;
D O I
10.1016/S0925-2312(97)00091-X
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The connections of the alternative model for mixture of experts (ME) to the normalized radial basis function (NRBF) nets and extended normalized RBF (ENRBF) nets are established, and the well-known expectation-maximization (EM) algorithm for maximum likelihood learning is suggested to the two types of RBF nets. This new learning technique determines the parameters of the input layer (including the covariance matrices or so-called receptive fields) and the parameters of the output layer in the RBF nets globally, instead of separately training the input layer by the K-means algorithm and the output layer by the least-squares learning as done in most of the existing RBF learning methods. In addition, coordinated competitive learning (CCL) and adaptive algorithms are proposed to approximate the EM algorithm for considerably speeding up the learning of the original and alternative ME models as well as the NRBF and ENRBF nets. Furthermore, the two ME models are linked to the recent proposed Bayesian Ying-Yang (BYY) learning system and theory such that not only the architecture of ME and RBF nets is shown to be more preferred than multilayer architecture, but also a new model selection criterion has been obtained to determine the number of experts and basis functions. A number of experiments are made on the prediction of foreign exchange rate and trading investment as well as piecewise nonlinear regression and piecewise line fitting. As shown in these experiments, the EM algorithm for NRBF nets and ENRBF nets obviously outperforms the conventional RBF learning technique, CCL speeds up the learning considerably with only a slight sacrifice on performance accuracy, the adaptive algorithm gives significant improvements on financial predication and trading investment, as well as that the proposed criterion can select the number of basis functions successfully. In addition, the ENRBF net and the alternative ME model are also shown to be able to implement curve fitting and detection. (C) 1998 Elsevier Science B.V. All rights reserved.
引用
收藏
页码:223 / 257
页数:35
相关论文
共 39 条
[11]   Adaptive Mixtures of Local Experts [J].
Jacobs, Robert A. ;
Jordan, Michael I. ;
Nowlan, Steven J. ;
Hinton, Geoffrey E. .
NEURAL COMPUTATION, 1991, 3 (01) :79-87
[12]  
JONES RD, 1991, P IJCNN91 SEATTL, V2, P473
[13]   Convergence results for the EM approach to mixtures of experts architectures [J].
Jordan, MI ;
Xu, L .
NEURAL NETWORKS, 1995, 8 (09) :1409-1431
[14]   HIERARCHICAL MIXTURES OF EXPERTS AND THE EM ALGORITHM [J].
JORDAN, MI ;
JACOBS, RA .
NEURAL COMPUTATION, 1994, 6 (02) :181-214
[15]   PROBABILISTIC AND NONPROBABILISTIC HOUGH TRANSFORMS - OVERVIEW AND COMPARISONS [J].
KALVIAINEN, H ;
HIRVONEN, P ;
XU, L ;
OJA, E .
IMAGE AND VISION COMPUTING, 1995, 13 (04) :239-252
[16]  
KARDIRKAMANATHA.V, 1991, ADV NEURAL INFORMATI, V3, P721
[17]  
Lei Xu, 1992, International Journal of Neural Systems, V3, P253, DOI 10.1142/S0129065792000218
[18]  
MEL BW, 1991, ADV NEURAL INFORMATI, V3, P757
[19]   Fast Learning in Networks of Locally-Tuned Processing Units [J].
Moody, John ;
Darken, Christian J. .
NEURAL COMPUTATION, 1989, 1 (02) :281-294
[20]  
NOWLAN SJ, 1990, CRGTR902 U TOR DEP C