Edge emotion recognition: applying fast Fourier transform on speech Mel spectrograms to classify emotion on a Raspberry Pi for near real-time analytics

Author:

de Andrade Dominik Esteves1,Buchkremer Rüdiger1

Affiliation:

1. FOM University of Applied Sciences

Abstract

Abstract Many people and machines are inherently unable to interpret socio-affective cues such as tone of voice. Thoughtful adoption of intelligent technologies may improve the conversation. Since direct communication often occurs via edge devices, where an additional network connection is often not guaranteed, we now describe a real-time processing method that captures and evaluates emotions in a speech via a terminal device such as the Raspberry Pi computer. In this article, we also present the current state of research on speech emotional recognition. We examine audio files from five important emotional speech databases and visualize them in situ with dB-scaled Mel spectrograms using TensorFlow and Matplotlib. Audio files are transformed using the fast Fourier transform method to generate spectrograms. For classification, a support vector machine kernel and a CNN with transfer learning are selected. The accuracy of this classification is 70% and 77%, respectively, a good value related to the execution of the algorithms on an edge device instead of on a server. On a Raspberry Pi, it took less than one second to evaluate pure emotion in speech using machine learning and the corresponding visualization, suggesting the speaker's emotional state.

Publisher

Research Square Platform LLC

Reference65 articles.

1. Abadi M, Agarwal A, Barham P et al (2019) Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv 2016. arXiv preprint arXiv:1603.04467

2. Mobile Edge Computing: A Survey;Abbas N;IEEE Internet Things J,2018

3. Evolution of Artificial Intelligence Programming Languages - a Systematic Literature Review;Adetiba E;J Comput Sci,2021

4. Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers;Akçay MB;Speech Commun,2020

5. Amiriparian S, Gerczuk M, Ottl S et al (2018) Bag-of-Deep-Features: Noise-Robust Deep Feature Representations for Audio Analysis. In: 2018 International Joint Conference on Neural Networks (IJCNN). IEEE, pp 1–7

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3