Abstract
Emotion recognition is a very challenging research field due to its complexity, as individual differences in cognitive–emotional cues involve a wide variety of ways, including language, expressions, and speech. If we use video as the input, we can acquire a plethora of data for analyzing human emotions. In this research, we use features derived from separately pretrained self-supervised learning models to combine text, audio (speech), and visual data modalities. The fusion of features and representation is the biggest challenge in multimodal emotion classification research. Because of the large dimensionality of self-supervised learning characteristics, we present a unique transformer and attention-based fusion method for incorporating multimodal self-supervised learning features that achieved an accuracy of 86.40% for multimodal emotion classification.
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference43 articles.
1. An Active Learning Paradigm for Online Audio-Visual Emotion Recognition;Kansizoglou;IEEE Trans. Affect. Comput.,2019
2. Yoon, S., Byun, S., and Jung, K. (2018, January 18–21). Multimodal Speech Emotion Recognition Using Audio and Text. Proceedings of the IEEE Spoken Language Technology Workshop (SLT), Athens, Greece.
3. Han, Z., Zhao, H., and Wang, R. (2019, January 27–29). Transfer Learning for Speech Emotion Recognition. Proceedings of the 2019 IEEE 5th Intl Conference on Big Data Security on Cloud (BigDataSecurity), IEEE Intl Conference on High Performance and Smart Computing, (HPSC) and IEEE Intl Conference on Intelligent Data and Security (IDS), Washington, DC, USA.
4. Ezzeldin, M., ElShaer, A., Wisdom, S., and Mishra, T. (2019). Transfer learning from sound representations for anger detection in speech. arXiv.
5. Nagarajan, B., and Oruganti, V.R.M. (2018). Deep net features for complex emotion recognition. arXiv.
Cited by
15 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献