LEARNING COMPLEX, EXTENDED SEQUENCES USING THE PRINCIPLE OF HISTORY COMPRESSION

被引:214
作者
SCHMIDHUBER, J [1 ]
机构
[1] UNIV COLORADO,DEPT COMP SCI,BOULDER,CO 80309
关键词
D O I
10.1162/neco.1992.4.2.234
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous neural network learning algorithms for sequence processing are computationally expensive and perform poorly when it comes to long time lags. This paper first introduces a simple principle for reducing the descriptions of event sequences without loss of information. A consequence of this principle is that only unexpected inputs can be relevant. This insight leads to the construction of neural architectures that learn to "divide and conquer" by recursively decomposing sequences. I describe two architectures. The first functions as a self-organizing multilevel hierarchy of recurrent networks. The second, involving only two recurrent networks, tries to collapse a multilevel predictor hierarchy into a single recurrent net. Experiments show that the system can require less computation per time step and many fewer training sequences than conventional training algorithms for recurrent nets.
引用
收藏
页码:234 / 242
页数:9
相关论文
共 12 条
[1]  
Hochreiter J., 1991, THESIS TU MUNCHEN
[2]  
MIYATA Y, 1988, 10TH P ANN C COGN SC, P223
[3]  
MOZER MC, 1990, CUCS49590 U COL TECH
[4]  
MYERS C, 1990, LEARNING DELAYED REI
[5]  
RING M, 1991, MACHINE LEARNING, P343
[6]  
ROBINSON AJ, 1987, CUEDFINFENGTR1 CAMBR
[7]  
ROHWER R, 1989, P JDISTRIBUTED ADAPT
[8]  
SCHMIDHUBER J, 1991, ARTIFICIAL NEURAL NETWORKS, VOLS 1 AND 2, P909
[9]  
Schmidhuber J., 1991, FKI14791 TU MUNCH I
[10]  
SCHMIDHUBER JH, 1991, FKI14891 TU MUNCH I