Artificial intelligence in music production: prospects for preparing future masters of musical art for creative activity in a modern professional environment

Authors

DOI:

https://doi.org/10.24195/artstudies.2025-3.13

Keywords:

artificial intelligence, music production, digital audio workstations (DAW), deep learning models (RNN, VAE, GAN, Transformer), training of future Masters of Musical Art, technological literacy

Abstract

The article is devoted to the problems of integrating artificial intelligence (AI) technologies into the digital music production cycle and its significance in the context of professional training of future Masters of Musical Art. The aim of the study is to trace the evolution of AI models from early connectionist and rule-based systems to modern transformer-based architectures, to characterise the factors influencing their applicability in professional workflows for generating musical material, and to determine their advantages and limitations in terms of preparing future Masters of Musical Art for successful creative activity in the modern professional environment.Research methods included retrospective analysis of key stages in the development of AI models, comparative analysis of recurrent neural networks, variational autoencoders, generative adversarial networks, and transformer-based systems, as well as a synthesis of literature using interdisciplinary sources in the fields of computer music, cognitive science, and music pedagogy.The results showed that AI in music production has evolved in line with a sequential process of developing new generations of models: from early symbolic systems to deep generative architectures and multimodal text-to-music conversion technologies. With each stage of model improvement, music producers gained new tools for creativity, but significant limitations and shortcomings that remain to this day necessitate the training of professionals to critically evaluate and adapt generated material integrated into digital audio workstations (DAW). The main prospects for preparing future Masters of Musical Art for creative activity in a modern professional environment are the need to acquire technological literacy, adaptability in working with generated material, and the ability to interact with AI in the format of joint creative and critical awareness of technological and artistic limitations. The prospect for further research is to determine the essence and functions of specific skills of future Masters of Musical Art to successfully engage in creative activity in a modern professional environment, implementing a cycle of music production using AI technologies.

References

1. Agostinelli, A., Denk, T. I., Borsos, Z., Engel, J., Verzetti, M., Caillon, A., ... & Frank, C. (2023). Musiclm: Generating music from text. arXiv preprint arXiv:2301.11325. https://doi.org/10.48550/arXiv.2301.11325

2. Civit, M., Civit-Masot, J., Cuadrado, F., & Escalona, M. J. (2022). A systematic review of artificial intelligence- based music generation: Scope, applications, and future trends. Expert Systems with Applications, 209, 118190. https://doi.org/10.1016/j.eswa.2022.118190

3. Collins, M. (2014). In the Box Music Production: Advanced Tools and Techniques for Pro Tools (1 ed.). Routledge. https://doi.org/10.4324/9780203066362

4. Dhariwal, P., Jun, H., Payne, C., Kim, J. W., Radford, A., & Sutskever, I. (2020). Jukebox: A generative model for music. arXiv:2005.00341. Retrieved from: https://assets.pubpub.org/2gnzbcnd/11608661311181.pdf

5. Eck, D., & Schmidhuber, J. (2002, August). Learning the long-term structure of the blues. In International Conference on Artificial Neural Networks (pp. 284–289). Berlin, Heidelberg: Springer Berlin Heidelberg. https://doi.org/10.1007/3-540-46084-5_47

6. Giuliani, L., De Filippo, A., & Borghesi, A. (2023). Towards Intelligent Music Production: A Sample-based Approach. In Ceur workshop proceedings (Vol. 3519, pp. 50–59). CEUR-WS. Retrieved from: https://hdl.handle.net/11585/952308

7. Goodfellow, I. J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., ... & Bengio, Y. (2014). Generative Adversarial Nets. Advances in Neural Information Processing Systems (NeurIPS), 27. https://doi.org/10.48550/arXiv.1406.2661

8. Hatem, G. (2023). The Role of Artificial Intelligence in Music Production: A Look at Amper Music’s Innovative Approach. Linkedin. Retrieved from: https://surl.lu/bffllx

9. Hiller, L. A., & Isaacson, L. M. (1979). Experimental Music; Composition with an electronic computer. Greenwood Publishing Group Inc. Retrieved from: https://dn790000.ca.archive.org/0/items/ experimentalmusi00hill/experimentalmusi00hill.pdf

10. Hochreiter, S., & Schmidhuber, J. (1997). Long Short-Term Memory. Neural Computation, 9 (8), 1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735

11. Huang, C. Z. A., Vaswani, A., Uszkoreit, J., Shazeer, N., Simon, I., Hawthorne, C., ... & Eck, D. (2018). Music transformer: Generating Music with Long- Term Structure. ICLR 2019 Conference Paper. Retrieved from: https://openreview.net/pdf?id=rJe4ShAcF7

12. Kingma, D. P., & Welling, M. (2013). Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114. https://doi.org/10.48550/arXiv.1312.6114

13. LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521 (7553), 436–444. https://doi.org/10.1038/nature14539

14. Lipton, Z. C., Berkowitz, J., & Elkan, C. (2015). A Critical Review of Recurrent Neural Networks for Sequence Learning. arXiv:1506.00019. https://doi.org/ 10.48550/arXiv.1506.00019

15. Liu, J. (2024). Expressive MIDI-format Piano Performance Generation. arXiv preprint arXiv:2408.00900. https://doi.org/10.48550/arXiv.2408.00900

16. Mitra, R., & Zualkernan, I. (2025). Music generation using deep learning and generative AI: a systematic review. IEEE Access, 13, 18079–18106. https://doi.org/10.1109/ACCESS.2025.3531798

17. Mozer, M. C. (1994). Neural Network Music Composition by Prediction: Exploring the Benefits of Psychoacoustic Constraints and Multi-scale Processing. Connection Science, 6 (2–3), 247–280. https://doi.org/ 10.1080/09540099408915726

18. MuseNet. (2019). OpenAI. Retrieved from: https://openai.com/index/musenet/?utm_source

19. Ovcharenko, N., Merezhko, Y., Moskva, O., Neboga, O., & Kosiak, L. (2021). Technological competence of future music teachers: diagnostic and formative methods. Amazonia Investiga, 10 (48), 238–247. https://doi.org/10.34069/AI/2021.48.12.25

20. Paleti, N. C. (2024). Positional Encoding Explained: A Deep Dive into Transformer PE. Medium. Retrieved from: https://medium.com/thedeephub/ positional-encoding-explained-a-deep-dive-into-transformer-pe-65cfe8cfe10b

21. Quick, D. (2016). Learning production probabilities for musical grammars. Journal of New Music Research, 45 (4), 295–313. https://doi.org/10.10 80/09298215.2016.1228680

22. Rastrygina, A. M. (2020). Training of future specialists-musicians on the principles of interdisciplinary models of free arts. Academic Notes. Series: Pedagogical Sciences, (190), 38–42. https://doi.org/10.36550/2415-7988-2020-1-190-38-42

23. Roberts, A., Engel, J., Raffel, C., Hawthorne, C., & Eck, D. (2018a). A hierarchical latent vector model for learning long-term structure in music. In International conference on machine learning (pp. 4364–4373). PMLR. Retrieved from: https://proceedings.mlr.press/ v80/roberts18a.html

24. Roberts, A., Engel, J., Raffel, C., Simon, I., & Hawthorne, C. (2018b). MusicVAE: Creating a palette for musical scores with machine learning. Magenta. Retrieved from: https://magenta.withgoogle.com/music-vae

25. Simpson, J., & Groff, B. (2023). The 5 Best AI Music Generators of 2023. Benjamin Groff. Retrieved from: https://www.benjamingroff.com/top-ai-music- generators/

26. Sterne, J., & Razlogova, E. (2021). Tuning sound for infrastructures: artificial intelligence, automation, and the cultural politics of audio mastering. Cultural Studies, 35 (4–5), 750–770. https://doi.org/10.1080/09502386. 2021.1895247

27. Tiwari, V. (2025). Attention Is All You Need – A Deep Dive into the Revolutionary Transformer Architecture. Towards AI. Retrieved from: https://surli.cc/ijvsla

28. Todd, P. M. (1989). A Connectionist Approach to Algorithmic Composition. Computer Music Journal, 13(4), 27–43. https://doi.org/10.2307/3679551

29. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2017). Attention Is All You Need. 31st Conference on Neural Information Processing Systems (NIPS 2017). Long Beach, CA, USA. https://doi.org/10.48550/arXiv.1706.03762

30. Wei, L., Yu, Y., Qin, Y., & Zhang, S. (2025). From Tools to Creators: A Review on the Development and Application of Artificial Intelligence Music Generation. Information, 16 (8), 656. https://doi.org/10.3390/info16080656

31. Yang, L. C., Chou, S. Y., & Yang, Y. H. (2017). MidiNet: A convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847. https://doi.org/10.48550/arXiv.1703.10847

32. Zhao, Y., Yang, M., Lin, Y., Zhang, X., Shi, F., Wang, Z., Ding, J., & Ning, H. (2025). AI-Enabled Text-to-Music Generation: A Comprehensive Review of Methods, Frameworks, and Future Directions. Electronics, 14 (6), 1197. https://doi.org/10.3390/electronics14061197

Downloads

Published

2025-10-14

Issue

Section

SECTION 1. CURRENT ISSUES OF ART PEDAGOGY