Author:
Song Yiying,Qu Yukun,Xu Shan,Liu Jia
Abstract
AbstractDeep convolutional neural networks (DCNN) nowadays can match and even outperform human performance in challenging complex tasks. However, it remains unknown whether DCNNs achieve human-like performance through human-like processes; that is, do DCNNs use similar internal representations to achieve the task as humans? Here we applied a reverse-correlation method to reconstruct the internal representations when DCNNs and human observers classified genders of faces. We found that human observers and a DCNN pre-trained for face identification, VGG-Face, showed high similarity between their “classification images” in gender classification, suggesting similar critical information utilized in this task. Further analyses showed that the similarity of the representations was mainly observed at low spatial frequencies, which are critical for gender classification in human studies. Importantly, the prior task experience, which the VGG-Face was pre-trained for processing faces at the subordinate level (i.e., identification) as humans do, seemed necessary for such representational similarity, because AlexNet, a DCNN pre-trained to process objects at the basic level (i.e., categorization), succeeded in gender classification but relied on a completely different representation. In sum, although DCNNs and humans rely on different sets of hardware to process faces, they can use a similar representation, possibly from similar prior task experiences, to achieve the same computation goal. Therefore, our study provides the first empirical evidence supporting the hypothesis of implementation-independent representation.
Publisher
Cold Spring Harbor Laboratory
Reference23 articles.
1. One weird trick for parallelizing convolutional neural networks;arXiv preprint,2014
2. Krizhevsky, A. , Sutskever, I. & Hinton, G.E. Imagenet classification with deep convolutional neural networks. in Advances in neural information processing systems 1097–1105 (2012).
3. Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014).
4. Parkhi, O.M. , Vedaldi, A. & Zisserman, A. Deep face recognition. (2015).
5. Taigman, Y. , Yang, M. , Ranzato, M.A. & Wolf, L. Deepface: Closing the gap to human-level performance in face verification. in Proceedings of the IEEE conference on computer vision and pattern recognition 1701–1708 (2014).
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献