Affiliation:
1. School of Economics, Business and Accounting, University of São Paulo, São Paulo 05508-900, Brazil
2. Faculty of Business Administration, Accountability and Economics, Federal University of Goiás, Goiânia 74690-900, Brazil
Abstract
Language models based on deep learning showed promising results for artistic generation purposes, including musical generation. However, the evaluation of symbolic musical generation models is mostly based on low-level mathematical metrics (e.g., the result of the loss function) due to the inherent difficulty in measuring the musical quality of a given performance due to the subjective nature of music. This work sought to measure and evaluate musical excerpts generated by deep learning models from a human perspective, limited to the scope of classical piano music generation. In this assessment, a population of 117 people performed blind tests with musical excerpts of human composition and musical excerpts generated using artificial intelligence models, including the models PerformanceRNN, Music Transformer, MuseNet and a custom model based on GRUs. The experiments demonstrated that musical excerpts generated using models based on the Transformer neural network architecture obtained the greatest receptivity within the tested population, surpassing the results of human compositions. In addition, the experiments also demonstrated that people with greater musical sensitivity and musical experience were more able to identify the compositional origin of the excerpts heard.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference28 articles.
1. Music, language, and composition;Adorno;Music. Q.,1993
2. Walder, C. (2016, January 16–18). Modeling symbolic music: Beyond the piano roll. Proceedings of the Asian Conference on Machine Learning, Hamilton, New Zealand.
3. Patel, A.D. (2010). Music, Language, and the Brain, Oxford University Press.
4. Graves, A. (2013). Generating sequences with recurrent neural networks. arXiv.
5. Mathews, A., Xie, L., and He, X. (2016, January 12–17). Senticap: Generating image descriptions with sentiments. Proceedings of the AAAI Con-ference on Artificial Intelligence, Phoenix, AZ, USA. No. 1.
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献