Abstract
In music information retrieval (MIR), beat tracking is one of the most fundamental tasks. To obtain this critical component from rhythmic music signals, a previous beat tracking system of hidden Markov model (HMM) with a recurrent neural network (RNN) observer was developed. Although the frequency of music beat is quite stable, existing HMM based methods do not take this feature into account. Accordingly, most of hidden states in these HMM-based methods are redundant, which is a disadvantage for time efficiency. In this paper, we proposed an efficient HMM using hidden states by exploiting the frequency contents of the neural network’s observation with Fourier transform, which extremely reduces the computational complexity. Observers that previous works used, such as bi-directional recurrent neural network (Bi-RNN) and temporal convolutional network (TCN), cannot perceive the frequency of music beat. To obtain more reliable frequencies from music, a periodic recurrent neural network (PRNN) based on attention mechanism is proposed as well, which is used as the observer in HMM. Experimental results on open source music datasets, such as GTZAN, Hainsworth, SMC, and Ballroom, show that our efficient HMM with PRNN is competitive to the state-of-the-art methods and has lower computational cost.
Funder
National Key Research and Development program of China
Fundamental Research Funds for the Central Universities
Subject
Electrical and Electronic Engineering,Computer Networks and Communications,Hardware and Architecture,Signal Processing,Control and Systems Engineering
Reference57 articles.
1. Neural tracking of the musical beat is enhanced by low-frequency sounds;Lenc;Proc. Natl. Acad. Sci. USA,2018
2. Music auto-tagging using deep Recurrent Neural Networks;Song;Neurocomputing,2018
3. Semantic Tagging of Singing Voices in Popular Music Recordings;Kim;IEEE/ACM Trans. Audio Speech Lang. Process.,2020
4. Music auto-tagging using scattering transform and convolutional neural network with self-attention;Song;Appl. Soft Comput.,2020
5. Wu, W., Han, F., Song, G., and Wang, Z. (December, January 30). Music genre classification using independent recurrent neural network. Proceedings of the 2018 Chinese Automation Congress (CAC), Xi’an, China.
Cited by
6 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献