Spontaneous Facial Behavior Analysis using Deep Transformer Based Framework for Child–Computer Interaction

Author:

Qayyum Abdul1,Razzak Imran2,Tanveer M.3,Mazher Moona4

Affiliation:

1. Department of Computer Science and Engineering, Université de Bourgogne, France,

2. School of Computer Science and Engineering, University of New South Wales, Sydney, Australia

3. Department of Mathematics, Indian Institute of Technology Indore, Simrol, Indore, 453552, India

4. Department of Computer and Mathematics, University Rovira i Virgili, Tarragona, Spain

Abstract

Abstract: A fascinating challenge in robotics-human interaction is imitating the emotion recognition capability of humans to robots with the aim to make human-robotics interaction natural, genuine and intuitive. To achieve the natural interaction in affective robots, human-machine interfaces, and autonomous vehicles, understanding our attitudes and opinions is very important, and it provides a practical and feasible path to realize the connection between machine and human. Multimodal interface that includes voice along with facial expression can manifest a large range of nuanced emotions compared to purely textual interfaces and provide a great value to improve the intelligence level of effective communication. Interfaces that fail to manifest or ignore user emotions may significantly impact the performance and risk being perceived as cold, socially inept, untrustworthy, and incompetent. To equip a child well for life, we need to help our children identify their feelings, manage them well, and express their needs in healthy, respectful, and direct ways. Early identification of emotional deficits can help to prevent low social functioning in children. In this work, we analyzed the child’s spontaneous behavior using multimodal facial expression and voice signal presenting multimodal transformer-based last feature fusion for facial behavior analysis in children to extract contextualized representations from RGB video sequence and Hematoxylin and eosin video sequence and then using these representations followed by pairwise concatenations of contextualized representations using cross-feature fusion technique to predict users emotions. To validate the performance of the proposed framework, we have performed experiments with the different pairwise concatenations of contextualized representations that showed significantly better performance than state of the art method. Besides, we perform t-distributed stochastic neighbor embedding visualization to visualize the discriminative feature in lower dimension space and probability density estimation to visualize the prediction capability of our proposed model.

Publisher

Association for Computing Machinery (ACM)

Subject

Computer Networks and Communications,Hardware and Architecture

Reference28 articles.

1. iAware: A Real-Time Emotional Biofeedback System Based on Physiological Signals

2. Hyper-Parameter Optimization for Emotion Detection using Physiological Signals

3. Amani Albraikan , Diana  P Tobón , and Abdulmotaleb El Saddik . 2018. Toward user-independent emotion recognition using physiological signals . IEEE sensors Journal 19, 19 ( 2018 ), 8402–8412. Amani Albraikan, Diana P Tobón, and Abdulmotaleb El Saddik. 2018. Toward user-independent emotion recognition using physiological signals. IEEE sensors Journal 19, 19 (2018), 8402–8412.

4. Russell Beale and Christian Peter . 2008. The role of affect and emotion in HCI . In Affect and emotion in human-computer interaction . Springer , 1–11. Russell Beale and Christian Peter. 2008. The role of affect and emotion in HCI. In Affect and emotion in human-computer interaction. Springer, 1–11.

5. Deformation-Based 3D Facial Expression Representation

Cited by 2 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Multi Fine-Grained Fusion Network for Depression Detection;ACM Transactions on Multimedia Computing, Communications, and Applications;2024-06-29

2. Deep Recurrent Regression with a Heatmap Coupling Module for Facial Landmarks Detection;Cognitive Computation;2022-10-27

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3