RBF nets, mixture experts, and Bayesian Ying-Yang learning

被引:76
作者
Xu, L [1 ]
机构
[1] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, NT, Peoples R China
关键词
radial basis function network; mixture of experts; Bayesian Ying-Yang learning; model selection; coordinated competitive learning; fast computation; adaptive algorithm; financial time series; curve fitting;
D O I
10.1016/S0925-2312(97)00091-X
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The connections of the alternative model for mixture of experts (ME) to the normalized radial basis function (NRBF) nets and extended normalized RBF (ENRBF) nets are established, and the well-known expectation-maximization (EM) algorithm for maximum likelihood learning is suggested to the two types of RBF nets. This new learning technique determines the parameters of the input layer (including the covariance matrices or so-called receptive fields) and the parameters of the output layer in the RBF nets globally, instead of separately training the input layer by the K-means algorithm and the output layer by the least-squares learning as done in most of the existing RBF learning methods. In addition, coordinated competitive learning (CCL) and adaptive algorithms are proposed to approximate the EM algorithm for considerably speeding up the learning of the original and alternative ME models as well as the NRBF and ENRBF nets. Furthermore, the two ME models are linked to the recent proposed Bayesian Ying-Yang (BYY) learning system and theory such that not only the architecture of ME and RBF nets is shown to be more preferred than multilayer architecture, but also a new model selection criterion has been obtained to determine the number of experts and basis functions. A number of experiments are made on the prediction of foreign exchange rate and trading investment as well as piecewise nonlinear regression and piecewise line fitting. As shown in these experiments, the EM algorithm for NRBF nets and ENRBF nets obviously outperforms the conventional RBF learning technique, CCL speeds up the learning considerably with only a slight sacrifice on performance accuracy, the adaptive algorithm gives significant improvements on financial predication and trading investment, as well as that the proposed criterion can select the number of basis functions successfully. In addition, the ENRBF net and the alternative ME model are also shown to be able to implement curve fitting and detection. (C) 1998 Elsevier Science B.V. All rights reserved.
引用
收藏
页码:223 / 257
页数:35
相关论文
共 39 条
[1]  
[Anonymous], 1995, Proceedings of International Conference on Neural Information Processing
[2]  
[Anonymous], 1982, Pattern recognition: A statistical approach
[3]  
[Anonymous], 1987, ALGORITHMS APPROXIMA
[4]  
BOTROS SM, 1991, ADV NEURAL INFORMATI, V3, P707
[5]  
Broomhead D. S., 1988, Complex Systems, V2, P321
[6]   ORTHOGONAL LEAST-SQUARES LEARNING ALGORITHM FOR RADIAL BASIS FUNCTION NETWORKS [J].
CHEN, S ;
COWAN, CFN ;
GRANT, PM .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1991, 2 (02) :302-309
[7]  
Cheung YM, 1996, PROCEEDINGS OF THE IEEE/IAFE 1996 CONFERENCE ON COMPUTATIONAL INTELLIGENCE FOR FINANCIAL ENGINEERING (CIFER), P141, DOI 10.1109/CIFER.1996.501838
[8]  
Devroye L., 1987, A course in density estimation
[9]  
Hart P.E., 1973, Pattern recognition and scene analysis
[10]   Layered Neural Networks with Gaussian Hidden Units as Universal Approximations [J].
Hartman, Eric J. ;
Keeler, James D. ;
Kowalski, Jacek M. .
NEURAL COMPUTATION, 1990, 2 (02) :210-215