Study on emotion recognition bias in different regional groups

Author:

Lukac Martin,Zhambulova Gulnaz,Abdiyeva Kamila,Lewis Michael

Abstract

AbstractHuman-machine communication can be substantially enhanced by the inclusion of high-quality real-time recognition of spontaneous human emotional expressions. However, successful recognition of such expressions can be negatively impacted by factors such as sudden variations of lighting, or intentional obfuscation. Reliable recognition can be more substantively impeded due to the observation that the presentation and meaning of emotional expressions can vary significantly based on the culture of the expressor and the environment within which the emotions are expressed. As an example, an emotion recognition model trained on a regionally-specific database collected from North America might fail to recognize standard emotional expressions from another region, such as East Asia. To address the problem of regional and cultural bias in emotion recognition from facial expressions, we propose a meta-model that fuses multiple emotional cues and features. The proposed approach integrates image features, action level units, micro-expressions and macro-expressions into a multi-cues emotion model (MCAM). Each of the facial attributes incorporated into the model represents a specific category: fine-grained content-independent features, facial muscle movements, short-term facial expressions and high-level facial expressions. The results of the proposed meta-classifier (MCAM) approach show that a) the successful classification of regional facial expressions is based on non-sympathetic features b) learning the emotional facial expressions of some regional groups can confound the successful recognition of emotional expressions of other regional groups unless it is done from scratch and c) the identification of certain facial cues and features of the data-sets that serve to preclude the design of the perfect unbiased classifier. As a result of these observations we posit that to learn certain regional emotional expressions, other regional expressions first have to be “forgotten”.

Publisher

Springer Science and Business Media LLC

Subject

Multidisciplinary

Reference47 articles.

1. Friesen, E. & Ekman, P. Facial action coding system: A technique for the measurement of facial movement. Palo Alto 3(2), 5 (1978).

2. Huang, C., Chen, G., Yu, H., Bao, Y. & Zhao, L. Speech emotion recognition under white noise, Arch. Acoust., 38(4), (2013). https://acoustics.ippt.pan.pl/index.php/aa/article/view/308

3. Huang, C. et al. Practical speech emotion recognition based on online learning: From acted data to elicited data. Math. Probl. Eng. 2013, 265819. https://doi.org/10.1155/2013/265819 (2013).

4. Mehrabian, A. Silent Messages (Wadsworth Pub. Co, 1971).

5. Mehrabian, A. ‘silent messages’ - a wealth of information about nonverbal communication (body language) (2009).

Cited by 2 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3