Abstract
Abstract
In recent years, human-computer interaction systems are gradually entering our lives. As one of the key technologies in human-computer interaction systems, Speech Emotion Recognition(SER) technology can accurately identify emotions and help machines better understand users’ intentions to improve the quality of human-computer interaction, which has received a lot of attention from researchers at home and abroad. With the successful application of deep learning in the fields of image recognition and speech recognition, scholars have started to try to use it in SER and have proposed many deep learning-based SER algorithms. In this paper, we conducted an in-depth study of these algorithms and found that they have problems such as too simple feature extraction methods, low utilization of human-designed features, high model complexity, and low accuracy of recognizing specific emotions. For the data processing, we quadrupled the RAVDESS dataset using additive Gaussian white noise (AWGN) for a total of 5760 audio samples. For the network structure, we build two parallel convolutional neural networks (CNN) to extract spatial features and a transformer encoder network to extract temporal features, classifying emotions from one of 8 classes. Taking advantage of CNN’s advantages in spatial feature representation and sequence encoding conversion, I obtained an accuracy of 80.46% on the hold-out test set of the RAVDESS data set.
Subject
General Physics and Astronomy
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Disruptive situation detection on public transport through speech emotion recognition;Intelligent Systems with Applications;2024-03
2. Emotion Recognition: Enhancing Human-Computer Interaction;2023 Eleventh International Conference on Intelligent Computing and Information Systems (ICICIS);2023-11-21
3. An Ensemble Transformer Model For Speech Emotion Recognition;2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT);2023-07-06
4. Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information;Applied Sciences;2023-02-08
5. Speech Emotion Classification: A Survey of the State-of-the-Art;Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering;2023