Facial Expression Recognition Methods in the Wild Based on Fusion Feature of Attention Mechanism and LBP
Author:
Liao Jun123, Lin Yuanchang13, Ma Tengyun13, He Songxiying3, Liu Xiaofang13, He Guotian13
Affiliation:
1. Chongqing Institute of Green Intelligent Technology, Chinese Academy of Sciences, Chongqing 400714, China 2. College of Mechanical Engineering, Chongqing University of Technology, Chongqing 400054, China 3. Chongqing Key Laboratory of Artificial Intelligence and Service Robot Control Technology, Chongqing Institute of Green Intelligent Technology, Chinese Academy of Sciences, Chongqing 400714, China
Abstract
Facial expression methods play a vital role in human–computer interaction and other fields, but there are factors such as occlusion, illumination, and pose changes in wild facial recognition, as well as category imbalances between different datasets, that result in large variations in recognition rates and low accuracy rates for different categories of facial expression datasets. This study introduces RCL-Net, a method of recognizing wild facial expressions that is based on an attention mechanism and LBP feature fusion. The structure consists of two main branches, namely the ResNet-CBAM residual attention branch and the local binary feature (LBP) extraction branch (RCL-Net). First, by merging the residual network and hybrid attention mechanism, the residual attention network is presented to emphasize the local detail feature information of facial expressions; the significant characteristics of facial expressions are retrieved from both channel and spatial dimensions to build the residual attention classification model. Second, we present a locally improved residual network attention model. LBP features are introduced into the facial expression feature extraction stage in order to extract texture information on expression photographs in order to emphasize facial feature information and enhance the recognition accuracy of the model. Lastly, experimental validation is performed using the FER2013, FERPLUS, CK+, and RAF-DB datasets, and the experimental results demonstrate that the proposed method has superior generalization capability and robustness in the laboratory-controlled environment and field environment compared to the most recent experimental methods.
Funder
National Key R&D Program Chinese Academy of Sciences “Light of the West” Talent Training Introduction Program Cooperation projects between Chongqing universities Chinese Academy of Sciences Chongqing technology innovation and application development special
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference57 articles.
1. Deep Facial Expression Recognition: A Survey;Li;IEEE Trans. Affective Comput.,2022 2. Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., and Matthews, I. (2010, January 13–18). The Extended Cohn-Kanade Dataset (CK+): A Complete Dataset for Action Unit and Emotion-Specified Expression. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops, San Francisco, CA, USA. 3. Lyons, M., Akamatsu, S., Kamachi, M., and Gyoba, J. (1998, January 14–16). Coding Facial Expressions with Gabor Wavelets. Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition, Nara, Japan. 4. Valstar, M.F., and Pantic, M. (2010, January 29). Induced Disgust, Happiness and Surprise: An Addition to the MMI Facial Expression Database. Proceedings of the 3rd International Workshop on Emotion, Paris, France. 5. Facial Expression Recognition from Near-Infrared Videos;Zhao;Image Vis. Comput.,2011
Cited by
15 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|