Slow feature analysis: Unsupervised learning of invariances

被引:901
作者
Wiskott, L [1 ]
Sejnowski, TJ
机构
[1] Salk Inst Biol Studies, Computat Neurobiol Lab, San Diego, CA 92168 USA
[2] Inst Adv Studies, D-14193 Berlin, Germany
[3] Humboldt Univ, Innovat Kolleg Theoret Biol, D-10115 Berlin, Germany
[4] Salk Inst Biol Studies, Howard Hughes Med Inst, La Jolla, CA 92037 USA
[5] Univ Calif San Diego, Dept Biol, La Jolla, CA 92037 USA
关键词
D O I
10.1162/089976602317318938
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Invariant features of temporally varying signals are useful for analysis and classification. Slow feature analysis (SFA) is a new method for learning invariant or slowly varying features from a vectorial input signal. It is based on a nonlinear expansion of the input signal and application of principal component analysis to this expanded signal and its time derivative. It is guaranteed to find the optimal solution within a family of functions directly and can learn to extract a large number of decor-related features, which are ordered by their degree of invariance. SFA can be applied hierarchically to process high-dimensional input signals and extract complex features. SFA is applied first to complex cell tuning properties based on simple cell output, including disparity and motion. Then more complicated input-output functions are learned by repeated application of SFA. Finally, a hierarchical network of SFA modules is presented as a simple model of the visual system. The same unstructured network can learn translation, size, rotation, contrast, or, to a lesser degree, illumination invariance for one-dimensional objects, depending on only the training stimulus. Surprisingly, only a few training objects suffice to achieve good generalization to new objects. The generated representation is suitable for object recognition. Performance degrades if the network is trained to learn multiple invariances simultaneously.
引用
收藏
页码:715 / 770
页数:56
相关论文
共 30 条
[1]  
[Anonymous], SHAPE SHADING
[2]  
BARROW HG, 1992, ARTIFICIAL NEURAL NE, V2, P881
[3]   Learning viewpoint-invariant face representations from visual experience in an attractor network [J].
Bartlett, MS ;
Sejnowski, TJ .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1998, 9 (03) :399-417
[4]  
Beck Nathaniel, 1996, POLIT ANAL, V6, P1, DOI [DOI 10.1093/PAN/6.1.1, 10.1093/pan/6.1.1]
[5]   SELF-ORGANIZING NEURAL NETWORK THAT DISCOVERS SURFACES IN RANDOM-DOT STEREOGRAMS [J].
BECKER, S ;
HINTON, GE .
NATURE, 1992, 355 (6356) :161-163
[7]  
BECKER S, 1995, BACKPROPAGATION THEO, P313
[8]  
BECKER S, 1993, ADV NEURAL INFORMATI, V5, P361
[9]   AN INFORMATION MAXIMIZATION APPROACH TO BLIND SEPARATION AND BLIND DECONVOLUTION [J].
BELL, AJ ;
SEJNOWSKI, TJ .
NEURAL COMPUTATION, 1995, 7 (06) :1129-1159
[10]  
Bishop C. M., 1995, NEURAL NETWORKS PATT