Applying Recurrent Neural Networks with integrated Attention Mechanism and Transformer Model for Automated Music Generation

Authors

  • Aditya Kumar Kalinga Institute of Industrial Technology, India
  • Ankita Lal Kalinga Institute of Industrial Technology, India

Keywords:

Music Generation, Recurrent Neural Network (RNN), Attention Mechanism, Transformer Model, Self-Similarity Matrix (SSM), Beat Level Segmentation

Abstract

One of the key applications of artificial intelligence has come out in the form of automated music generation, that is a hybrid of creativity and computational models. The main purpose of this research is to develop and understand the blend of advanced Recurrent Neural Networks(RNNs) architecture with attention mechanisms to improve the limitations of ongoing RNNs in capitalizing long term dependencies. The main targets of the model are on relevant segments of input sequences,ensuring enhanced consistency and structural strength in the generated music industry. By comparison of enhanced RNN with a Transformer based model known for its exceptional capacity to model long-range dependencies through self-assessment. Additionally to filter rhythmic exactness and style, a beat-level segmentation method is implemented into this process. The structural composition of the generated outputs is examined using a Self-Similarity Matrix (SSM), which balances between reiteration and diversity.

References

[1] Jiang, R., & Mou, X. (2024), "The Analysis of Multi-Track Music Generation With Deep Learning Models in Music Production Process,” IEEE Access, 12, 110322-110330.

[2] Mienye, Ibomoiye Domor, Theo G. Swart, and George Obaido. "Recurrent neural networks: A comprehensive review of architectures, variants, and applications." 2 Inf., 15, 517.

[3] Ilhan, F., Karaahmetoglu, O., Balaban, I., & Kozat, S.S. (2020). “Markovian RNN: An Adaptive Time Series Prediction Network With HMM-Based Switching for Nonstationary Environments. IEEE Transactions on Neural Networks and Learning Systems,” 34, 715-728.

[4] Sajad, S., Dharshika, S., & Meleet, M. (2021). “Music Generation for Novices Using Recurrent Neural Network (RNN). 2021 International Conference on Innovative Computing,” Intelligent Communication and Smart Electrical Systems (ICSES), 1-6.

[5] Gunawan, A.A., Iman, A.P., & Suhartono, D. (2020). “Automatic Music Generator Using Recurrent Neural Network.” Int. J. Comput. Intell. Syst., 13, 645-654.

[6] Nasrullah, Z., & Zhao, Y. (2019). “Music Artist Classification with Convolutional Recurrent Neural Networks.” 2019 International Joint Conference on Neural Networks (IJCNN), 1-8.

[7] Zhang, Y., & Li, Z. (2021). “Automatic Synthesis Technology of Music Teaching Melodies Based on Recurrent Neural Network.” Sci. Program., 2021, 1704995:1-1704995:10.

[8] Hsu, J., & Chang, S. (2021). “Generating Music Transition by Using a Transformer-Based Model. Electronics.”

[9] Jin, C., Wang, T., Liu, S., Yun, T., Li, J., Li, X., & Lui, S. (2020). “A Transformer-Based Model for Multi-Track Music Generation.” Int. J. Multim. Data Eng. Manag., 11, 36-54.

[10] Huang, Y., & Yang, Y. (2020). Pop Music Transformer: Beat-based Modeling and Generation of Expressive Pop Piano Compositions. Proceedings of the 28th ACM International Conference on Multimedia.

Downloads

Published

2024-10-31

How to Cite

Aditya Kumar, & Ankita Lal. (2024). Applying Recurrent Neural Networks with integrated Attention Mechanism and Transformer Model for Automated Music Generation. International Journal on Smart & Sustainable Intelligent Computing, 1(2), 58–69. Retrieved from https://submissions.adroidjournals.com/index.php/ijssic/article/view/34

Issue

Section

Research Articles