Context-free and context-sensitive dynamics in recurrent neural networks

被引:33
作者
Bodén, M [1 ]
Wiles, J [1 ]
机构
[1] Univ Queensland, Dept Comp Sci & Elect Engn, Brisbane, Qld 4072, Australia
关键词
context-free grammar; context-sensitive grammar; dynamical system language; learning; recurrent neural network; recursive structure;
D O I
10.1080/095400900750060122
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continuous-valued recurrent neural networks can learn mechanisms for processing context-free languages. The dynamics of such networks is usually based on damped oscillation around fixed points in state space and requires that the dynamical components are arranged in certain ways. It is shown that qualitatively similar dynamics with similar constraints hold for a(n)b(n)c(n), a context-sensitive language. The additional difficulty with a(n)b(n)c(n), compared with the context-free language a(n)b(n), consists of 'counting up' and 'counting down' letters simultaneously. The network solution is to oscillate in two principal dimensions, one for counting up and one for counting down. This study focuses on the dynamics employed by the sequential cascaded network, in contrast to the simple recurrent network, and the use of backpropagation through time. Found solutions generalize well beyond training data, however, learning is not reliable. The contribution of this study lies in demonstrating how the dynamics in recurrent neural networks that process context-free languages can also be employed in processing some context-sensitive languages (traditionally thought of as requiring additional computation resources). This continuity of mechanism between language classes contributes to our understanding of neural networks in modelling language learning and processing.
引用
收藏
页码:197 / 210
页数:14
相关论文
共 28 条
  • [1] [Anonymous], 1974, Differential Equations, Dynamical Systems, and Linear Algebra
  • [2] [Anonymous], PROCEEDINGS OF THE S
  • [3] [Anonymous], 1994, P C ADV NEUR INF PRO
  • [4] Analysis of dynamical recognizers
    Blair, AD
    Pollack, JB
    [J]. NEURAL COMPUTATION, 1997, 9 (05) : 1127 - 1142
  • [5] Bodén M, 1999, IEE CONF PUBL, P359, DOI 10.1049/cp:19991135
  • [6] The dynamics of discrete-time computation, with application to recurrent neural networks and finite state machine extraction
    Casey, M
    [J]. NEURAL COMPUTATION, 1996, 8 (06) : 1135 - 1178
  • [7] CHALUP S, 1999, P 6 INT C NEUR INF P, P508
  • [8] Toward a connectionist model of recursion in human linguistic performance
    Christiansen, MH
    Chater, N
    [J]. COGNITIVE SCIENCE, 1999, 23 (02) : 157 - 205
  • [9] Finite State Automata and Simple Recurrent Networks
    Cleeremans, Axel
    Servan-Schreiber, David
    McClelland, James L.
    [J]. NEURAL COMPUTATION, 1989, 1 (03) : 372 - 381
  • [10] DAS S, 1993, ADV NEURAL INFORMATI, V5, P65