Artificial intelligence in music production: prospects for preparing future masters of musical art for creative activity in a modern professional environment
DOI:
https://doi.org/10.24195/artstudies.2025-3.13Keywords:
artificial intelligence, music production, digital audio workstations (DAW), deep learning models (RNN, VAE, GAN, Transformer), training of future Masters of Musical Art, technological literacyAbstract
The article is devoted to the problems of integrating artificial intelligence (AI) technologies into the digital music production cycle and its significance in the context of professional training of future Masters of Musical Art. The aim of the study is to trace the evolution of AI models from early connectionist and rule-based systems to modern transformer-based architectures, to characterise the factors influencing their applicability in professional workflows for generating musical material, and to determine their advantages and limitations in terms of preparing future Masters of Musical Art for successful creative activity in the modern professional environment.Research methods included retrospective analysis of key stages in the development of AI models, comparative analysis of recurrent neural networks, variational autoencoders, generative adversarial networks, and transformer-based systems, as well as a synthesis of literature using interdisciplinary sources in the fields of computer music, cognitive science, and music pedagogy.The results showed that AI in music production has evolved in line with a sequential process of developing new generations of models: from early symbolic systems to deep generative architectures and multimodal text-to-music conversion technologies. With each stage of model improvement, music producers gained new tools for creativity, but significant limitations and shortcomings that remain to this day necessitate the training of professionals to critically evaluate and adapt generated material integrated into digital audio workstations (DAW). The main prospects for preparing future Masters of Musical Art for creative activity in a modern professional environment are the need to acquire technological literacy, adaptability in working with generated material, and the ability to interact with AI in the format of joint creative and critical awareness of technological and artistic limitations. The prospect for further research is to determine the essence and functions of specific skills of future Masters of Musical Art to successfully engage in creative activity in a modern professional environment, implementing a cycle of music production using AI technologies.
References
1. Agostinelli, A., Denk, T. I., Borsos, Z., Engel, J., Verzetti, M., Caillon, A., ... & Frank, C. (2023). Musiclm: Generating music from text. arXiv preprint arXiv:2301.11325. https://doi.org/10.48550/arXiv.2301.11325
2. Civit, M., Civit-Masot, J., Cuadrado, F., & Escalona, M. J. (2022). A systematic review of artificial intelligence- based music generation: Scope, applications, and future trends. Expert Systems with Applications, 209, 118190. https://doi.org/10.1016/j.eswa.2022.118190
3. Collins, M. (2014). In the Box Music Production: Advanced Tools and Techniques for Pro Tools (1 ed.). Routledge. https://doi.org/10.4324/9780203066362
4. Dhariwal, P., Jun, H., Payne, C., Kim, J. W., Radford, A., & Sutskever, I. (2020). Jukebox: A generative model for music. arXiv:2005.00341. Retrieved from: https://assets.pubpub.org/2gnzbcnd/11608661311181.pdf
5. Eck, D., & Schmidhuber, J. (2002, August). Learning the long-term structure of the blues. In International Conference on Artificial Neural Networks (pp. 284–289). Berlin, Heidelberg: Springer Berlin Heidelberg. https://doi.org/10.1007/3-540-46084-5_47
6. Giuliani, L., De Filippo, A., & Borghesi, A. (2023). Towards Intelligent Music Production: A Sample-based Approach. In Ceur workshop proceedings (Vol. 3519, pp. 50–59). CEUR-WS. Retrieved from: https://hdl.handle.net/11585/952308
7. Goodfellow, I. J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., ... & Bengio, Y. (2014). Generative Adversarial Nets. Advances in Neural Information Processing Systems (NeurIPS), 27. https://doi.org/10.48550/arXiv.1406.2661
8. Hatem, G. (2023). The Role of Artificial Intelligence in Music Production: A Look at Amper Music’s Innovative Approach. Linkedin. Retrieved from: https://surl.lu/bffllx
9. Hiller, L. A., & Isaacson, L. M. (1979). Experimental Music; Composition with an electronic computer. Greenwood Publishing Group Inc. Retrieved from: https://dn790000.ca.archive.org/0/items/ experimentalmusi00hill/experimentalmusi00hill.pdf
10. Hochreiter, S., & Schmidhuber, J. (1997). Long Short-Term Memory. Neural Computation, 9 (8), 1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
11. Huang, C. Z. A., Vaswani, A., Uszkoreit, J., Shazeer, N., Simon, I., Hawthorne, C., ... & Eck, D. (2018). Music transformer: Generating Music with Long- Term Structure. ICLR 2019 Conference Paper. Retrieved from: https://openreview.net/pdf?id=rJe4ShAcF7
12. Kingma, D. P., & Welling, M. (2013). Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114. https://doi.org/10.48550/arXiv.1312.6114
13. LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521 (7553), 436–444. https://doi.org/10.1038/nature14539
14. Lipton, Z. C., Berkowitz, J., & Elkan, C. (2015). A Critical Review of Recurrent Neural Networks for Sequence Learning. arXiv:1506.00019. https://doi.org/ 10.48550/arXiv.1506.00019
15. Liu, J. (2024). Expressive MIDI-format Piano Performance Generation. arXiv preprint arXiv:2408.00900. https://doi.org/10.48550/arXiv.2408.00900
16. Mitra, R., & Zualkernan, I. (2025). Music generation using deep learning and generative AI: a systematic review. IEEE Access, 13, 18079–18106. https://doi.org/10.1109/ACCESS.2025.3531798
17. Mozer, M. C. (1994). Neural Network Music Composition by Prediction: Exploring the Benefits of Psychoacoustic Constraints and Multi-scale Processing. Connection Science, 6 (2–3), 247–280. https://doi.org/ 10.1080/09540099408915726
18. MuseNet. (2019). OpenAI. Retrieved from: https://openai.com/index/musenet/?utm_source
19. Ovcharenko, N., Merezhko, Y., Moskva, O., Neboga, O., & Kosiak, L. (2021). Technological competence of future music teachers: diagnostic and formative methods. Amazonia Investiga, 10 (48), 238–247. https://doi.org/10.34069/AI/2021.48.12.25
20. Paleti, N. C. (2024). Positional Encoding Explained: A Deep Dive into Transformer PE. Medium. Retrieved from: https://medium.com/thedeephub/ positional-encoding-explained-a-deep-dive-into-transformer-pe-65cfe8cfe10b
21. Quick, D. (2016). Learning production probabilities for musical grammars. Journal of New Music Research, 45 (4), 295–313. https://doi.org/10.10 80/09298215.2016.1228680
22. Rastrygina, A. M. (2020). Training of future specialists-musicians on the principles of interdisciplinary models of free arts. Academic Notes. Series: Pedagogical Sciences, (190), 38–42. https://doi.org/10.36550/2415-7988-2020-1-190-38-42
23. Roberts, A., Engel, J., Raffel, C., Hawthorne, C., & Eck, D. (2018a). A hierarchical latent vector model for learning long-term structure in music. In International conference on machine learning (pp. 4364–4373). PMLR. Retrieved from: https://proceedings.mlr.press/ v80/roberts18a.html
24. Roberts, A., Engel, J., Raffel, C., Simon, I., & Hawthorne, C. (2018b). MusicVAE: Creating a palette for musical scores with machine learning. Magenta. Retrieved from: https://magenta.withgoogle.com/music-vae
25. Simpson, J., & Groff, B. (2023). The 5 Best AI Music Generators of 2023. Benjamin Groff. Retrieved from: https://www.benjamingroff.com/top-ai-music- generators/
26. Sterne, J., & Razlogova, E. (2021). Tuning sound for infrastructures: artificial intelligence, automation, and the cultural politics of audio mastering. Cultural Studies, 35 (4–5), 750–770. https://doi.org/10.1080/09502386. 2021.1895247
27. Tiwari, V. (2025). Attention Is All You Need – A Deep Dive into the Revolutionary Transformer Architecture. Towards AI. Retrieved from: https://surli.cc/ijvsla
28. Todd, P. M. (1989). A Connectionist Approach to Algorithmic Composition. Computer Music Journal, 13(4), 27–43. https://doi.org/10.2307/3679551
29. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2017). Attention Is All You Need. 31st Conference on Neural Information Processing Systems (NIPS 2017). Long Beach, CA, USA. https://doi.org/10.48550/arXiv.1706.03762
30. Wei, L., Yu, Y., Qin, Y., & Zhang, S. (2025). From Tools to Creators: A Review on the Development and Application of Artificial Intelligence Music Generation. Information, 16 (8), 656. https://doi.org/10.3390/info16080656
31. Yang, L. C., Chou, S. Y., & Yang, Y. H. (2017). MidiNet: A convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847. https://doi.org/10.48550/arXiv.1703.10847
32. Zhao, Y., Yang, M., Lin, Y., Zhang, X., Shi, F., Wang, Z., Ding, J., & Ning, H. (2025). AI-Enabled Text-to-Music Generation: A Comprehensive Review of Methods, Frameworks, and Future Directions. Electronics, 14 (6), 1197. https://doi.org/10.3390/electronics14061197







