Folk melody generation based on CNN-BiGRU and Self-Attention | IEEE Conference Publication | IEEE Xplore

Folk melody generation based on CNN-BiGRU and Self-Attention


Abstract:

Using various deep learning networks to generate music is a research hotspot in the field of human intelligence. Due to the limitation of network structure in the existin...Show More

Abstract:

Using various deep learning networks to generate music is a research hotspot in the field of human intelligence. Due to the limitation of network structure in the existing melody generation models, the quality of melody generation is poor. In this paper, for folk melody, we propose a melody generation network based on CNN-BiGRU and Self-Attention. First, the processed data is input into the CNN module consisting of one-dimensional convolution layer and pooling layer to extract local features. Then, the features extracted by CNN module are constructed as time series and is input into BiGRU network to extract global features. Finally, Self-Attention assigns different weights to the features of BiGRU output to highlight important feature information during note generation. Experimental results show that the prediction accuracy of the proposed model is improved by 3.61 % compared with BiGUR, 5.26% compared with CNN, and 6.53% compared with CNN-BiGRU. Meanwhile, the model presented in this paper also achieves improvement in other evaluation measures.
Date of Conference: 27-29 May 2022
Date Added to IEEE Xplore: 17 August 2022
ISBN Information:
Conference Location: Shenzhen, China

Funding Agency:

References is not available for this document.

I. Introduction

Automatic melody generation means that melody can be automatically generated quickly by learning the theoretical rules of composition through computer[I]. Automatic melody generation can assist the composer to create music. At the same time, it can help music lovers compose their own melodies and lower the threshold of melody creation[2].

Select All
1.
S Jain and A Pandian, "A survey on automatic music generation", International Journal of Engineering & Technology, vol. 7, no. 2.8, pp. 677, 2018.
2.
A. Natsiou and S. O'Leary, "Audio representations for deep learning in sound synthesis: A review", IEEE/ACS 18th International Conference on Computer Systems and Applications (AICCSA), pp. 1-8, 2021.
3.
D. Arora, M. Garg and M. Gupta, "Diving deep in Deep Convolutional Neural Network", 2020 2nd International Conference on Advances in Computing Communication Control and Networking (ICACCCN), pp. 749-751, 2020.
4.
A. Maduskar, A. Ladukar, S. Gore and N. Patwari, "Music Generation using Deep Generative Modelling", 2020 International Conference on Convergence to Digital World - Quo Vadis (ICCDW), pp. 1-4, 2020.
5.
X. Zhuang, T. Jiang, S. -Y. Chou, B. Wu, P. Hu and S. Lui, "Litesing: Towards Fast Lightweight and Expressive Singing Voice Synthesis", ICASSP 2021–2021 IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp. 7078-7082, 2021.
6.
W. Zhao, Y. Zhou, Y. Tie and Y. Zhao, "Recurrent Neural Network for MIDI Music Emotion Classification", 2018 IEEE 3rd Advanced Information Technology Electronic and Automation Control Conference (IAEAC), pp. 2596-2600, 2018.
7.
JS Wang and J Zhou, "Polyphonic Music Generation Method Based on charRNN", Computer Engineering, vol. 45, pp. 249-255, 2019.
8.
J. Wu, C. Hu, Y. Wang, X. Hu and J. Zhu, "A Hierarchical Recurrent Neural Network for Symbolic Melody Generation", IEEE Transactions on Cybernetics, vol. 50, no. 6, pp. 2749-2757, 2020.
9.
M Farzaneh and R M Toroghi, Music Generation Using an Interactive Evolutionary Algorithm, pp. 207-217, 2020.
10.
S. Chikkamath and N. S R, "Melody generation using LSTM and BI-LSTM Network", 2021 International Conference on Computational Intelligence and Computing Applications (ICCICA), pp. 1-6, 2021.
11.
A Vaswani, N Shazeer and N Parmar, Attention Is All You Need, 2017.
12.
N Zhang and J Yan, Melody Structure Transfer Network: Generating Music with Separable Self-Attention, 2021.
13.
G Keerti, AN Vaishnavi and P Mukherjee, "Attentional networks for music generation", Multimedia Tools and Applications, vol. 81, no. 4, pp. 5179-5189, 2022.
14.
H. Li and F. Ren, "Computer music and traditional music under the background of “Internet + Music” – Take the film “Song of the Phoenix” for example", 2020 5th International Conference on Mechanical Control and Computer Engineering (ICMCCE), pp. 1458-1461, 2020.
15.
Q Fu, "Research on the Use of Computer Music in Modern Musical Composition", Journal of Physics: Conference Series, vol. 1820, no. 1, pp. 012153, 2021.
16.
Ping-Huan and Chiou-Jye H, "A high precision artitficial neural networks mode for short-term energy load forecasting", Energies, vol. 11, no. 1, pp. 213-226, 2018.
17.
Zhengyang Tang, "Melody generation algorithm based on LSTM network", Information and computer, vol. 14, pp. 62-64, 2021.
18.
R Kumar, A Biswas and P Roy, Melody Extraction from Music: A Comprehensive Study, pp. 141-155, 2020.

Contact IEEE to Subscribe

References

References is not available for this document.