UNIFIED FRAMEWORK FOR MLPS AND RBFNS - INTRODUCING CONIC SECTION FUNCTION NETWORKS

被引:29
作者
DORFFNER, G
机构
[1] Department of Medical Cybernetics and Artificial Intelligence, University of Vienna, Austrian Research Institute for Artificial Intelligence, Vienna
关键词
D O I
10.1080/01969729408902340
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multilayer perceptrons (MLPs) (Werbos, 1974; Rumelhart et al., 1986) and radial basis function networks (RBI;Ns) (Broomhead and Lowe, 1988; Moody and Darken, 1989) are probably the most widely used neural network models for practical applications. Whereas the former belong to a group of ''classical'' neural networks (whose weighted sums are loosely inspired by biology), the latter have risen only recently from an analogy to regression theory (Broomhead and Lowe, 1988). On first sight, the twomodels-except for being multilayer feedforward networks-do not seem to have much in common. On second thought, however, MLPs and RBFNs share a variety of features, worthy of viewing them in the same context and comparing them to each other with respect to their properties. Consequently, a few attempts on arriving at a unified picture of a class of feedforward networks-with MLPs and RBFNs as members-have been undertaken (Robinson et al., 1988; Maruyama et al., 1992; Dorffner, 1992). Most of these attempts have centered around the observation that the function of a neural network unit can be divided into a propagation rule (net input) and an activation or transfer function. The dot product (weighted sum) and the Euclidean distance are special cases of propagation rules, whereas the sigmoid and Gaussian function are examples of activation functions. This paper introduces a novel neural network model based on a more general conic section function as a propagation rule, containing hyperplane (straight line) and hypersphere (circle) as special cases, thus unifying the net inputs of MLPs and RBFNs with an easy-to-handle continuum in between. A new learning rule-complementing the existing methods of gradient descent in weight space and initialization-is introduced which enables the network to make a continuous decision between bounded and unbounded (infinite half-space) decision regions. The capabilities of conic section function networks (CSFNs) are illustrated with several examples and compared with existing approaches. CSFNs are viewed as a further step toward more efficient and optimal neural network solutions in practical applications.
引用
收藏
页码:511 / 554
页数:44
相关论文
共 32 条
[1]  
Improving upon standard pattern classification algorithms by implementing them as multilayer perceptrons, RSRE Memorandum, pp. 43-46, (1989)
[2]  
Bock H.H., Automalische Kiassifikation, (1974)
[3]  
Broomhead D.S., Lowe D., Multivariable functional interpolation and adaptive networks, Complex Syst, 2, pp. 321-355, (1988)
[4]  
Chen S., Cowan C.F.N., Orthogonal least squares learning algorithm for radial basis function networks, IEEE Trans. Neural Networks, 2, 2, pp. 302-309, (1991)
[5]  
Chui C.K., Li X., Mhaskar H.N., Neural Networks for Localized Approximation, (1992)
[6]  
Cybenko G., Approximation by superpositions of a sigmoidal function, Math Control Signals Syst, 2, pp. 303-314, (1989)
[7]  
Dcnoeux T., Lengelle R., Initializing backpropagation networks with prototypes, Neural Networks, 6, 3, pp. 351-364, (1993)
[8]  
Dorffner G., Euclid net—a multilayer feedforward network using the Euclidean distance as propagation rule, Artificial Neural Networks 2, 2, pp. 1633-1636, (1992)
[9]  
Porenta G.G., On using feedforward neural networks for clinical diagnostic tasks, In Artificial Intelligence in Medicine, (1994)
[10]  
Wiklicky G.H., Reanalysing similarity measures in neural networks and their practical consequences, Report Oefai-92-15, (1992)