Slow Feature Analysis: Unsupervised Learning of Invariances

Slow Feature Analysis: Unsupervised Learning of Invariances,10.1162/089976602317318938,Neural Computation,Laurenz Wiskott,Terrence J. Sejnowski

Slow Feature Analysis: Unsupervised Learning of Invariances   (Citations: 200)
BibTex | RIS | RefWorks Download
Invariant features of temporally varying signals are useful for analysis and classification. Slow feature analysis (SFA) is a new method for learn- ing invariant or slowly varying features from a vectorial input signal. It is based on a nonlinear expansion of the input signal and applica- tion of principal component analysis to this expanded signal and its time derivative. It is guaranteed to find the optimal solution within a family of functions directly and can learn to extract a large number of decor- related features, which are ordered by their degree of invariance. SFA can be applied hierarchically to process high-dimensional input signals and extract complex features. SFA is applied first to complex cell tuning properties based on simple cell output, including disparity and motion. Then more complicated input-output functions are learned by repeated application of SFA. Finally, a hierarchical network of SFA modules is presented as a simple model of the visual system. The same unstructured network can learn translation, size, rotation, contrast, or, to a lesser degree, illumination invariance for one-dimensional objects, depending on only the training stimulus. Surprisingly, only a few training objects suffice to achieve good generalization to new objects. The generated representation is suitable for object recognition. Performance degrades if the network is trained to learn multiple invariances simultaneously.
Journal: Neural Computation - NECO , vol. 14, no. 4, pp. 715-770, 2002
Cumulative Annual
View Publication
The following links allow you to view full publications. These links are maintained by other sources not affiliated with Microsoft Academic Search.
Sort by: