Recurrent nets that time and count

被引:401
作者
Gers, FA [1 ]
Schmidhuber, J [1 ]
机构
[1] IDSIA, CH-6900 Lugano, Switzerland
来源
IJCNN 2000: PROCEEDINGS OF THE IEEE-INNS-ENNS INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOL III | 2000年
关键词
D O I
10.1109/IJCNN.2000.861302
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
The size of the time intervals between events conveys information essential for numerous sequential tasks such as motor control and rhythm detection. While Hidden Markov Models tend to ignore this information, recurrent neural networks (RNNs) can in principle learn to make use of it. We focus on Long Short-Term Memory (LSTM) because it usually outperforms other RNNs. Surprisingly, LSTM augmented by "peephole connections" from its internal cells to its multiplicative gates can learn the fine distinction between sequences of spikes separated by either 50 or 49 discrete time steps, without the help of any short training exemplars. Without external resets or teacher forcing or loss of performance on tasks reported earlier, our LSTM variant also learns to generate very stable sequences of highly nonlinear, precisely timed spikes. This makes LSTM a promising approach for real-world tasks that require to time and count.
引用
收藏
页码:189 / 194
页数:6
相关论文
共 9 条
[1]
[Anonymous], PROCEEDINGS OF THE S
[2]
ADAPTIVE NEURAL OSCILLATOR USING CONTINUOUS-TIME BACK-PROPAGATION LEARNING [J].
DOYA, K ;
YOSHIZAWA, S .
NEURAL NETWORKS, 1989, 2 (05) :375-385
[3]
Fu-Sheng Tsung, 1995, Advances in Neural Information Processing Systems 7, P481
[4]
Gers FA, 1999, IEE CONF PUBL, P850, DOI [10.1162/089976600300015015, 10.1049/cp:19991218]
[5]
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[6]
GRADIENT CALCULATIONS FOR DYNAMIC RECURRENT NEURAL NETWORKS - A SURVEY [J].
PEARLMUTTER, BA .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1995, 6 (05) :1212-1228
[7]
Plaut DC., 1986, Experiments on Learning by Back Propagation
[8]
Rodriguez P, 1998, ADV NEUR IN, V10, P87
[9]
A Learning Algorithm for Continually Running Fully Recurrent Neural Networks [J].
Williams, Ronald J. ;
Zipser, David .
NEURAL COMPUTATION, 1989, 1 (02) :270-280