Abstract
AbstractReflectance, lighting, and geometry combine in complex ways to create images. How do we disentangle these to perceive individual properties, like surface glossiness? We suggest that brains disentangle properties by learning to model statistical structure in proximal images. To test this, we trained unsupervised generative neural networks on renderings of glossy surfaces and compared their representations with human gloss judgments. The networks spontaneously cluster images according to distal properties such as reflectance and illumination, despite receiving no explicit information about them. Intriguingly, the resulting representations also predict the specific patterns of ‘successes’ and ‘errors’ in human perception. Linearly decoding specular reflectance from the model’s internal code predicts human gloss perception better than ground truth, supervised networks, or control models, and predicts, on an image-by-image basis, illusions of gloss perception caused by interactions between material, shape, and lighting. Unsupervised learning may underlie many perceptual dimensions in vision, and beyond.
Publisher
Cold Spring Harbor Laboratory
Reference117 articles.
1. Adelson, E. H. Lightness Perception and Lightness Illusions. in The New Cognitive Neurosciences 339–351 (MIT Press, 2000).
2. Mid-level vision;Current Biology,2020
3. Anderson, B. L . The perceptual representation of transparency, lightness, and gloss. Handbook of Perceptual Organization 466–483 (2015).
4. Recovering intrinsic scene characteristics;Computer Vision Systems,1978
5. Material perception;Annual Review of Vision Science,2017
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献