Local linear perceptrons for classification

被引:45
作者
Alpaydin, E [1 ]
Jordan, MI [1 ]
机构
[1] MIT,DEPT BRAIN & COGNIT SCI,CAMBRIDGE,MA 02139
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1996年 / 7卷 / 03期
关键词
D O I
10.1109/72.501737
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A structure composed of local linear perceptrons for approximating global class discriminants is investigated, Such local linear models may be combined in a cooperative or competitive way, In the cooperative model, a weighted sum of the outputs of the local perceptrons is computed where the weight is a function of the distance between the input and the position of the local perceptron, In the competitive model, the cost function dictates a mixture model where only one of the local perceptrons give output, Learning of the local models' positions and the linear mappings they implement are coupled and both supervised, We show that this is preferrable to the uncoupled case where the positions are trained in an unsupervised manner before the separate, supervised training of mappings, We use goodness criteria based on the cross-entropy and give learning equations for both the cooperative and competitive cases, The coupled and uncoupled versions of cooperative and competitive approaches are compared among themselves and with multi-lager perceptrons of sigmoidal hidden units and radial basis functions (RBF's) of Gaussian units an the application of recognition of handwritten digits, The criteria of comparison are the generalization accuracy, learning time, and the number of free parameters, We conclude that even on such a high-dimensional problem, such local models are promising, They generalize much better than RBF's and use much less memory, When compared with multilayer perceptrons, we note that local models learn much faster and generalize as well and sometimes better with comparable number of parameters.
引用
收藏
页码:788 / 792
页数:5
相关论文
共 15 条
[1]  
[Anonymous], 1994, 5469 NISTIR
[2]   LOCAL LEARNING ALGORITHMS [J].
BOTTOU, L ;
VAPNIK, V .
NEURAL COMPUTATION, 1992, 4 (06) :888-900
[3]  
Bridle J. S., 1990, Neurocomputing, Algorithms, Architectures and Applications. Proceedings of the NATO Advanced Research Workshop, P227
[4]   Product Units: A Computationally Powerful and Biologically Plausible Extension to Backpropagation Networks [J].
Durbin, Richard ;
Rumelhart, David E. .
NEURAL COMPUTATION, 1989, 1 (01) :133-142
[5]  
Hampshire J. B., 1990, ADV NEURAL INFORMATI, V2, P203
[6]   Adaptive Mixtures of Local Experts [J].
Jacobs, Robert A. ;
Jordan, Michael I. ;
Nowlan, Steven J. ;
Hinton, Geoffrey E. .
NEURAL COMPUTATION, 1991, 3 (01) :79-87
[7]   HIERARCHICAL MIXTURES OF EXPERTS AND THE EM ALGORITHM [J].
JORDAN, MI ;
JACOBS, RA .
NEURAL COMPUTATION, 1994, 6 (02) :181-214
[8]   NEURAL-GAS NETWORK FOR VECTOR QUANTIZATION AND ITS APPLICATION TO TIME-SERIES PREDICTION [J].
MARTINETZ, TM ;
BERKOVICH, SG ;
SCHULTEN, KJ .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1993, 4 (04) :558-569
[9]   Fast Learning in Networks of Locally-Tuned Processing Units [J].
Moody, John ;
Darken, Christian J. .
NEURAL COMPUTATION, 1989, 1 (02) :281-294
[10]  
MURRAYSMITH R, 1994, THESIS U STRATHCLYDE