Affiliation:
1. Rajiv Gandhi College of Engineering Research and Technology, Chandrapur, Maharashtra, India
Abstract
This research paper presents a Speech Emotion Recognition (SER) system utilizing a Multilayer Perceptron (MLP) classifier and real-time audio analysis. The system records audio samples, extracts relevant features, and employs machine learning techniques to predict emotions in spoken language. The study focuses on the development of an intuitive Graphical User Interface (GUI) using the Kivy framework, providing a user-friendly platform for real-time emotion analysis. In shortly In this project, we attempt to detect underlying emotions such as (sad, happy, neutral, angry, disgust, surprised, fearful and calm) in recorded speech by analysing the acoustic features of the audio data of recordings and Created an application to implement the same on user input.
Reference15 articles.
1. Han, K., et al. "Speech Emotion Recognition Using Deep Neural Network and Extreme Learning Machine." Proceedings of the International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2014.
2. Badshah, A. M., et al. "Emotion Speech Recognition System for Smart Affective Services Based on Deep Functionality." IEEE Access, 2019.
3. Saito, Y., et al. "Statistical Parametric Speech Synthesis Incorporating Generative Adversarial Networks." IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2018.
4. Zhou, X., et al. "Deep Learning Based Affective Model for Speech Emotion Recognition." IEEE Transactions on Affective Computing, 2017.
5. Zhang, T., et al. "Speech Emotion Recognition with i-vector Feature and MN Model." IEEE Transactions on Multimedia, 2016.