Using machine learning analysis to interpret the relationship between music emotion and lyric features

Author:

Xu Liang1,Sun Zaoyi2,Wen Xin1,Huang Zhengxi1,Chao Chi-ju3,Xu Liuchang45

Affiliation:

1. Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, China

2. College of Education, Zhejiang University of Technology, Hangzhou, China

3. Department of Information Art and Design, Tsinghua University, Beijing, China

4. Zhejiang Provincial Key Laboratory of Forestry Intelligent Monitoring and Information Technology, Zhejiang A&F University, Hangzhou, China

5. College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou, China

Abstract

Melody and lyrics, reflecting two unique human cognitive abilities, are usually combined in music to convey emotions. Although psychologists and computer scientists have made considerable progress in revealing the association between musical structure and the perceived emotions of music, the features of lyrics are relatively less discussed. Using linguistic inquiry and word count (LIWC) technology to extract lyric features in 2,372 Chinese songs, this study investigated the effects of LIWC-based lyric features on the perceived arousal and valence of music. First, correlation analysis shows that, for example, the perceived arousal of music was positively correlated with the total number of lyric words and the mean number of words per sentence and was negatively correlated with the proportion of words related to the past and insight. The perceived valence of music was negatively correlated with the proportion of negative emotion words. Second, we used audio and lyric features as inputs to construct music emotion recognition (MER) models. The performance of random forest regressions reveals that, for the recognition models of perceived valence, adding lyric features can significantly improve the prediction effect of the model using audio features only; for the recognition models of perceived arousal, lyric features are almost useless. Finally, by calculating the feature importance to interpret the MER models, we observed that the audio features played a decisive role in the recognition models of both perceived arousal and perceived valence. Unlike the uselessness of the lyric features in the arousal recognition model, several lyric features, such as the usage frequency of words related to sadness, positive emotions, and tentativeness, played important roles in the valence recognition model.

Funder

The Research and Development Foundation of Zhejiang A&F University

The Open Research Fund of Zhejiang Provincial Key Laboratory of Resources and Environmental Information System

The Scientific Research Foundation of the Education Department of Zhejiang Province, China

Publisher

PeerJ

Subject

General Computer Science

Cited by 4 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Using Deep Learning to Recognize Therapeutic Effects of Music Based on Emotions;Sensors;2023-01-14

2. DeepLyric: Predicting Music Emotions through LSTM-GRU Hybrid Models with Regularization Techniques;Procedia Computer Science;2023

3. Multimodal Music Emotion Recognition based on WLDNN_GAN;2022 International Symposium on Advances in Informatics, Electronics and Education (ISAIEE);2022-12

4. Integrating Individual Factors to Construct Recognition Models of Consumer Fraud Victimization;International Journal of Environmental Research and Public Health;2022-01-01

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3