Multimodal Facial Emotion Recognition Using Improved Convolution Neural Networks Model
-
Published:2023-07-20
Issue:4
Volume:27
Page:710-719
-
ISSN:1883-8014
-
Container-title:Journal of Advanced Computational Intelligence and Intelligent Informatics
-
language:en
-
Short-container-title:JACIII
Author:
Udeh Chinonso Paschal123, Chen Luefeng123, Du Sheng123, Li Min123, Wu Min123
Affiliation:
1. School of Automation, China University of Geosciences, No.388 Lumo Road, Hongshan District, Wuhan 430074, China 2. Hubei Key Laboratory of Advanced Control and Intelligent Automation for Complex Systems, No.388 Lumo Road, Hongshan District, Wuhan 430074, China 3. Engineering Research Center of Intelligent Technology for Geo-Exploration, Ministry of Education, No.388 Lumo Road, Hongshan District, Wuhan 430074, China
Abstract
In the quest for human-robot interaction (HRI), leading to the development of emotion recognition, learning, and analysis capabilities, robotics plays a significant role in human perception, attention, decision-making, and social communication. However, the accurate recognition of emotions in HRI remains a challenge. This is due to the coexistence of multiple sources of information in utilizing multimodal facial expressions and head poses as multiple convolutional neural networks (CNN) and deep learning are combined. This research analyzes and improves the robustness of emotion recognition, and proposes a novel approach that optimizes traditional deep neural networks that fall into poor local optima when optimizing the weightings of the deep neural network using standard methods. The proposed approach adaptively finds the better weightings of the network, resulting in a hybrid genetic algorithm with stochastic gradient descent (HGASGD). This hybrid algorithm combines the inherent, implicit parallelism of the genetic algorithm with the better global optimization of stochastic gradient descent (SGD). An experiment shows the effectiveness of our proposed approach in providing complete emotion recognition through a combination of multimodal data, CNNs, and HGASGD, indicating that it represents a powerful tool in achieving interactions between humans and robotics. To validate and test the effectiveness of our proposed approach through experiments, the performance and reliability of our approach and two variants of HGASGD FER are compared using a large dataset of facial images. Our approach integrates multimodal information from facial expressions and head poses, enabling the system to recognize emotions better. The results show that CNN-HGASGD outperforms CNNs-SGD and other existing state-of-the-art methods in terms of FER.
Funder
National Natural Science Foundation of China Higher Education Discipline Innovation Project Fundamental Research Funds for the Central Universities, China University of Geosciences
Publisher
Fuji Technology Press Ltd.
Subject
Artificial Intelligence,Computer Vision and Pattern Recognition,Human-Computer Interaction
Reference45 articles.
1. F. Foroni and G. R. Semin, “Language that puts you in touch with your bodily feelings: The multimodal responsiveness of affective expressions,” Psychological Science, Vol.20, No.8, pp. 974-980, 2009. https://doi.org/10.1111/j.1467-9280.2009.02400.x 2. A. L. Thomaz and C. Breazeal, “Teachable robots: Understanding human teaching behavior to build more effective robot learners,” Artificial Intelligence, Vol.172, No.6-7, pp. 716-737, 2008. https://doi.org/10.1016/j.artint.2007.09.009 3. C. Korsmeyer and R. W. Picard, “Affective Computing,” Minds and Machines, Vol.9, pp. 443-447, 1999. https://doi.org/10.1023/A:1008329803271 4. L. Chen, M. Wu, M. Zhou, Z. Liu, J. She, and K. Hirota, “Dynamic emotion understanding in human-robot interaction based on two-layer fuzzy SVR-TS model,” IEEE Trans. on Systems, Man, and Cybernetics: Systems, Vol.50, No.2, pp. 490-501, 2020. https://doi.org/10.1109/TSMC.2017.2756447 5. F. Afza, M. A. Khan, M. Sharif, S. Kadry, G. Manogaran, T. Saba, I. Ashraf, and R. Damaševičius, “A framework of human action recognition using length control features fusion and weighted entropy-variances based feature selection,” Image and Vision Computing, Vol.106, Article No.104090, 2021. https://doi.org/10.1016/j.imavis.2020.104090
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|