|
[1] G. Brunner, A. Konrad, Y. Wang, and R. Wattenhofer. Midi-vae: Modeling dynamics and instrumentation of music with applications to style transfer. arXiv preprint arXiv:1809.07600, 2018. [2] P. Gomez and B. Danuser. Relationships between musical structure and psychophysiological measures of emotion. Emotion, 7(2):377, 2007. [3] J. Grekow. From content-based music emotion recognition to emotion maps of musical pieces, volume 747. Springer, 2018. [4] J. Grekow and T. Dimitrova-Grekow. Monophonic music generation with a given emotion using conditional variational autoencoder. IEEE Access, 9:129088–129101, 2021. [5] I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick, S. Mohamed, and A. Lerchner. beta-vae: Learning basic visual concepts with a constrained variational framework. 2016. [6] W.-Y. Hsiao, J.-Y. Liu, Y.-C. Yeh, and Y.-H. Yang. Compound word transformer: Learning to compose full-song music over dynamic directed hypergraphs. arXiv preprint arXiv:2101.02402, 2021. [7] W.-N. Hsu, Y. Zhang, R. J. Weiss, H. Zen, Y. Wu, Y. Wang, Y. Cao, Y. Jia, Z. Chen, J. Shen, et al. Hierarchical generative modeling for controllable speech synthesis. arXiv preprint arXiv:1810.07217, 2018. [8] Y.-S. Huang and Y.-H. Yang. Pop music transformer: Beat-based modeling and generation of expressive pop piano compositions, 2020. [9] H.-T. Hung, J. Ching, S. Doh, N. Kim, J. Nam, and Y.-H. Yang. Emopia: A multi-modal pop piano dataset for emotion recognition and emotion-based music generation. arXiv preprint arXiv:2108.01374, 2021. [10] Z. Jiang, Y. Zheng, H. Tan, B. Tang, and H. Zhou. Variational deep embedding: An unsupervised and generative approach to clustering, 2016. [11] L. Kawai, P. Esling, and T. Harada. Attributes-aware deep music transformation. In Proceedings of the 21st international society for music information retrieval conference, ismir, 2020. [12] N. S. Keskar, B. McCann, L. R. Varshney, C. Xiong, and R. Socher. Ctrl: A conditional transformer language model for controllable generation. arXiv preprint arXiv:1909.05858, 2019. [13] D. P. Kingma and M. Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013. [14] D. Makris, K. R. Agres, and D. Herremans. Generating lead sheets with affect: A novel conditional seq2seq framework. In 2021 International Joint Conference on Neural Networks (IJCNN), pages 1–8. IEEE, 2021. [15] F. C. A. E. F. S. N. G. Nathan Fradet, Jean-Pierre Briot. Miditok: A python package for midi file tokenization. In Extended Abstracts for the Late-Breaking Demo Session of the 22nd International Society for Music Information Retrieval Conference, 2021. [16] A. Pati and A. Lerch. Latent space regularization for explicit control of musical attributes. In ICML Machine Learning for Music Discovery Workshop (ML4MD), Extended Abstract, Long Beach, CA, USA, 2019. [17] J. A. Russell. A circumplex model of affect. Journal of personality and social psychology, 39(6):1161, 1980. [18] K. R. Scherer. What are emotions? and how can they be measured? Social science information, 44(4):695–729, 2005. [19] K. Sohn, H. Lee, and X. Yan. Learning structured output representation using deep conditional generative models. Advances in neural information processing systems, 28, 2015. [20] H. H. Tan and D. Herremans. Music fadernets: Controllable music generation based on high-level features via low-level feature modelling. In Proc. of the International Society for Music Information Retrieval Conference, 2020. [21] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. [22] S.-L. Wu and Y.-H. Yang. MuseMorphose: Full-song and fine-grained music style transfer with one Transformer VAE. arXiv preprint arXiv:2105.04090, 2021. [23] L.-C. Yang, S.-Y. Chou, and Y.-H. Yang. Midinet: A convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847, 2017. |