Abstract
This paper presents a theoretical perspective on modeling ventral stream processing by revisiting the computational abstraction of simple and complex cells. In parallel to David Marr's vision theory, we organize the new perspective into three levels. At the computational level, we abstract simple and complex cells into space partitioning and composition in a topological space based on the redundancy exploitation hypothesis of Horace Barlow. At the algorithmic level, we present a hierarchical extension of sparse coding by exploiting the manifold constraint in high-dimensional space (i.e., the blessing of dimensionality). The resulting over-parameterized models for object recognition differ from existing hierarchical models by disentangling the objectives of selectivity and invariance computation. It is possible to interpret our hierarchical construction as a computational implementation of cortically local subspace untangling for object recognition and face representation, which are closely related to exemplar-based and axis-based coding in the medial temporal lobe. At the implementation level, we briefly discuss two possible implementations based on asymmetric sparse autoencoders and divergent spiking neural networks.
Subject
Cellular and Molecular Neuroscience,Neuroscience (miscellaneous)
Reference98 articles.
1. On the Surprising Behavior of Distance Metrics in High Dimensional Space
2. K-svd: an algorithm for designing overcomplete dictionaries for sparse representation;Aharon;IEEE Transact. Signal Process,2006
3. “Approximate nearest neighbors and the fast johnson-lindenstrauss transform,”;Ailon;Proceedings of the Thirty-Eighth Annual ACM Symposium on Theory of Computing,2006
4. “A convergence theory for deep learning via over-parameterization,”;Allen-Zhu;International Conference on Machine Learning, vol. 97,2019
5. “On the optimization of deep networks: Implicit acceleration by overparameterization,”;Arora;International Conference on Machine Learning, vol. 80,2018