Abstract
AbstractThe ability of music to induce emotions has been arousing a lot of interest in recent years, especially due to the boom in music streaming platforms and the use of automatic music recommenders. Music Emotion Recognition approaches are based on combining multiple audio features extracted from digital audio samples and different machine learning techniques. In these approaches, neuroscience results on musical emotion perception are not considered. The main goal of this research is to facilitate the automatic subtitling of music. The authors approached the problem of automatic musical emotion detection in movie soundtracks considering these characteristics and using scientific musical databases, which have become a reference in neuroscience research. In the experiments, the Constant-Q-Transform spectrograms, the ones that best represent the relationships between musical tones from the point of view of human perception, are combined with Convolutional Neural Networks. Results show an efficient emotion classification model for 2-second musical audio fragments representative of intense basic feelings of happiness, sadness, and fear. Those emotions are the most interesting to be identified in the case of movie music captioning. The quality metrics have demonstrated that the results of the different models differ significantly and show no homogeneity. Finally, these results pave the way for an accessible and automatic captioning of music, which could automatically identify the emotional intent of the different segments of the movie soundtrack.
Funder
Ministerio de Ciencia, Innovación y Universidades
Comunidad de Madrid
Publisher
Springer Science and Business Media LLC
Reference40 articles.
1. AENOR. (2012) Norma UNE 153010 Subtitulado para personas sordas y personas con discapacidad auditiva [Norm UNE 153010. Subtitling for deaf and hearing-impaired persons]
2. Che X, Luo S, Yang H, Meinel C (2017) Automatic Lecture Subtitle Generation and How It Helps. In Proceedings - IEEE 17th International Conference on Advanced Learning Technologies, ICALT 2017, pages 34–38
3. Donnelly KJ (2005) The Spectre of Sound. British Film Institute, London
4. Thompson WF, Russo FA, Sinclair D (1994) Effects of underscoring on the perception of closure in filmed events. Psychomusicol J Res Music Cogn 13(1–2):9–27
5. Pehrs C, Deserno L, Bakels J H, Schlochtermeier L H, Kappelhoff H, Jacobs A M, Fritz T H, Koelsch S, Kuchinke L (2014) How music alters a kiss: Superior temporal gyrus controls fusiform-amygdalar effective connectivity. Soc Cognitive Affect Neurosci 9(11):1770–1778 11
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献