Learning algorithms based on linearization

被引:3
作者
Hahnloser, R [1 ]
机构
[1] ETH Honggerberg, Inst Theoret Phys, CH-8093 Zurich, Switzerland
[2] ETH Zurich, UNIZ, Inst Neuroinformat, CH-8006 Zurich, Switzerland
关键词
D O I
10.1088/0954-898X/9/3/006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
The aim of this article is to investigate:a mechanical description of learning. A framework for local and simple learning algorithms based on interpreting a neural network as a set of configuration constraints ia proposed. For any architectural design and learning task, unsupervised and supervised algorithms can be derived, optionally using unconstrained and hidden neurons. Unlike algorithms based on the gradient in weight space, the proposed tangential correlation (TC) algorithms move along the gradient in state space. This results in optimal scaling properties and simple expressions for the weight updates. The number of synapses is much larger than the number of neurons. A constraint for neural states does not impose a unique constraint for synaptic weights. Which weights to assign credit to can be selected from a parametrization of all weight changes equivalently satisfying the state constraints. At the heart of the parametrization are minimal weight changes. Two supervised algorithms (differing by their parametrizations) operating on a three-layer perceptron are compared with standard backpropagation. The successful training of fixed points of recurrent networks is demonstrated. The unsupervised learning of oscillations with variable frequencies is performed on standard and more sophisticated recurrent networks. The results presented here can be useful both for the analysis and for the synthesis of learning algorithms.
引用
收藏
页码:363 / 380
页数:18
相关论文
共 30 条
[1]
ALMEIDA LB, 1987, 1ST P IEEE INT C NEU, V2, P609
[2]
Natural gradient works efficiently in learning [J].
Amari, S .
NEURAL COMPUTATION, 1998, 10 (02) :251-276
[3]
QUANTITATIVE STUDY OF ATTRACTOR NEURAL NETWORK RETRIEVING AT LOW SPIKE RATES .1. SUBSTRATE SPIKES, RATES AND NEURONAL GAIN [J].
AMIT, DJ ;
TSODYKS, MV .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1991, 2 (03) :259-273
[4]
ARBIB MA, 1987, BRAIN MACHINES MATH, P63
[5]
TRAINING A 3-NODE NEURAL NETWORK IS NP-COMPLETE [J].
BLUM, AL ;
RIVEST, RL .
NEURAL NETWORKS, 1992, 5 (01) :117-127
[6]
NONSUPERVISED ADAPTIVE SIGNAL DETECTION + PATTERN RECOGNITION [J].
COOPER, DB ;
COOPER, PW .
INFORMATION AND CONTROL, 1964, 7 (03) :416-&
[7]
RECURRENT EXCITATION IN NEOCORTICAL CIRCUITS [J].
DOUGLAS, RJ ;
KOCH, C ;
MAHOWALD, M ;
MARTIN, KAC ;
SUAREZ, HH .
SCIENCE, 1995, 269 (5226) :981-985
[8]
Product Units: A Computationally Powerful and Biologically Plausible Extension to Backpropagation Networks [J].
Durbin, Richard ;
Rumelhart, David E. .
NEURAL COMPUTATION, 1989, 1 (01) :133-142
[9]
HIGH-FREQUENCY (60-90 HZ) OSCILLATIONS IN PRIMARY VISUAL-CORTEX OF AWAKE MONKEY [J].
ECKHORN, R ;
FRIEN, A ;
BAUER, R ;
WOELBERN, T ;
KEHR, H .
NEUROREPORT, 1993, 4 (03) :243-246
[10]
Propagation of activity-dependent synaptic depression in simple neural networks [J].
Fitzsimonds, RM ;
Song, HJ ;
Poo, MM .
NATURE, 1997, 388 (6641) :439-448