Author:
Korcsok Beáta,Faragó Tamás,Ferdinandy Bence,Miklósi Ádám,Korondi Péter,Gácsi Márta
Abstract
AbstractEmotionally expressive non-verbal vocalizations can play a major role in human-robot interactions. Humans can assess the intensity and emotional valence of animal vocalizations based on simple acoustic features such as call length and fundamental frequency. These simple encoding rules are suggested to be general across terrestrial vertebrates. To test the degree of this generalizability, our aim was to synthesize a set of artificial sounds by systematically changing the call length and fundamental frequency, and examine how emotional valence and intensity is attributed to them by humans. Based on sine wave sounds, we generated sound samples in seven categories by increasing complexity via incorporating different characteristics of animal vocalizations. We used an online questionnaire to measure the perceived emotional valence and intensity of the sounds in a two-dimensional model of emotions. The results show that sounds with low fundamental frequency and shorter call lengths were considered to have a more positive valence, and samples with high fundamental frequency were rated as more intense across all categories, regardless of the sound complexity. We conclude that applying the basic rules of vocal emotion encoding can be a good starting point for the development of novel non-verbal vocalizations for artificial agents.
Publisher
Springer Science and Business Media LLC
Reference72 articles.
1. Mavridis, N. A review of verbal and non-verbal human-robot interactive communication. Rob. Auton. Syst. 63, 22–35 (2015).
2. Bennewitz, M., Faber, F., Joho, D. & Behnke, S. Fritz - A Humanoid Communication Robot. In RO-MAN 2007 - The 16th IEEE International Symposium on Robot and Human Interactive Communication 1072–1077, https://doi.org/10.1109/ROMAN.2007.4415240 (IEEE, 2007).
3. Meena, R., Jokinen, K. & Wilcock, G. Integration of gestures and speech in human-robot interaction. 3rd IEEE Int. Conf. Cogn. Infocommunications, CogInfoCom 2012 - Proc. 673–678, https://doi.org/10.1109/CogInfoCom.2012.6421936 (2012).
4. Pellegrino, F., Coupé, C. & Marsico, E. Across-Language Perspective on Speech Information Rate. Language (Baltim). 87, 539–558 (2012).
5. Ekman, P. & Friesen, W. Unmasking the face: A guide to recognizing emotions from facial clues. (ISHK, 2003).
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献