|
[1] Y. Ren, J. He, X. Tan, T. Qin, Z. Zhao, and T.-Y. Liu, “Popmag: Pop mu- sic accompaniment generation,” in Proceedings of the 28th ACM International Conference on Multimedia, MM ’20, (New York, NY, USA), p. 1198–1206, Association for Computing Machinery, 2020. [2] H. Zhu, Q. Liu, N. J. Yuan, C. Qin, J. Li, K. Zhang, G. Zhou, F. Wei, Y. Xu, and E. Chen, “Xiaoice band: A melody and arrangement generation framework for pop music,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD ’18, (New York, NY, USA), p. 2837–2846, Association for Computing Machinery, 2018. [3] N. Jiang, S. Jin, Z. Duan, and C. Zhang, “Rl-duet: Online music accompa- niment generation using deep reinforcement learning,” in The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, pp. 710–718, AAAI Press, 2020. [4] I. Simon, D. Morris, and S. Basu, “Mysong: Automatic accompaniment gener- ation for vocal melodies,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI ’08, (New York, NY, USA), p. 725–734, Association for Computing Machinery, 2008. [5] C.-Z. A. Huang, A. Vaswani, J. Uszkoreit, I. Simon, C. Hawthorne, N. M. Shazeer, A. M. Dai, M. Hoffman, M. Dinculescu, and D. Eck, “Music trans- former: Generating music with long-term structure,” in ICLR, 2019. [6] A. Roberts, J. Engel, C. Raffel, C. Hawthorne, and D. Eck, “A hierarchical latent vector model for learning long-term structure in music,” in International Conference on Machine Learning (ICML), 2018. [7] P. L. Diéguez and V.-W. Soo, “Variational autoencoders for polyphonic music interpolation,” in 2020 International Conference on Technologies and Applica- tions of Artificial Intelligence (TAAI), pp. 56–61, 2020. [8] K. Choi, C. Hawthorne, I. Simon, M. Dinculescu, and J. Engel, “Encoding musical style with transformer autoencoders,” in ICML, 2020. [9] T. Borghuis, A. Tibo, S. Conforti, L. Canciello, L. Brusci, and P. Frasconi, “Off the beaten track: Using deep learning to interpolate between music genres,” CoRR, vol. abs/1804.09808, 2018. [10] T. Borghuis, A. Tibo, S. Conforti, L. Brusci, and P. Frasconi, “Full-band music genres interpolations with wasserstein autoencoders,” in Ital-IA, 2019. Work- shop AI for Media and Entertainment ; Conference date: 18-03-2019 Through 18-03-2019. [11] D. Berthelot, C. Raffel, A. Roy, and I. Goodfellow, “Understanding and im- proving interpolation in autoencoders via an adversarial regularizer,” in Inter- national Conference on Learning Representations, 2019. [12] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. u. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems (I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, eds.), vol. 30, Curran Associates, Inc., 2017. [13] I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural networks,” in Advances in Neural Information Processing Systems (Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Q. Weinberger, eds.), vol. 27, Curran Associates, Inc., 2014. [14] P. Todd and G. Loy, “A connectionist approach to algorithmic composition,” Computer Music Journal, vol. 13, pp. 173–194, 1989. [15] S. Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Com- putation, vol. 9, pp. 1735–1780, 11 1997. [16] D. Eck and J. Schmidhuber, “A first look at music composition using lstm recurrent neural networks,” tech. rep., 2002. [17] T. Jiang, Q. Xiao, and X. Yin, “Music generation using bidirectional recurrent network,” in 2019 IEEE 2nd International Conference on Electronics Technol- ogy (ICET), pp. 564–569, 2019. [18] S. Ji, J. Luo, and X. Yang, “A comprehensive survey on deep music gener- ation: Multi-level representations, algorithms, evaluations, and future direc- tions,” ArXiv, vol. abs/2011.06801, 2020. [19] KevinL, “Midi: Who we are.” https://www.midi.org/about; accessed 5-Dec- 2021. [20] Y.-S. Huang and Y.-H. Yang, “Pop music transformer: Beat-based modeling and generation of expressive pop piano compositions,” in Proceedings of the 28th ACM International Conference on Multimedia, MM ’20, (New York, NY, USA), p. 1180–1188, Association for Computing Machinery, 2020. [21] D. C. Corrêa and F. Rodrigues, “A survey on symbolic data-based music genre classification,” Expert Syst. Appl., vol. 60, pp. 190–210, 2016. [22] R. Hillewaere, B. Manderick, and D. Conklin, “Alignment methods for folk tune classification,” in Data Analysis, Machine Learning and Knowledge Discovery, pp. 369–377, Springer, 2014. [23] M. Zeng, X. Tan, R. Wang, Z. Ju, T. Qin, and T.-Y. Liu, “MusicBERT: Symbolic music understanding with large-scale pre-training,” in Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, (Online), pp. 791–800, Association for Computational Linguistics, Aug. 2021. [24] H. Liang, W. Lei, P. Y. Chan, Z. Yang, M. Sun, and T.-S. Chua, “Pirhdy: Learning pitch-, rhythm-, and dynamics-aware embeddings for symbolic mu- sic,” in Proceedings of the 28th ACM International Conference on Multimedia, pp. 574–582, 2020. [25] L. Li, R. Zhang, and Z. Wang, “Melodic phrase attention network for symbolic data-based music genre classification (student abstract),” in AAAI, 2021. [26] C. Liu, L. Feng, G. Liu, H. Wang, and S. Liu, “Bottom-up broadcast neural network for music genre classification,” Multim. Tools Appl., vol. 80, pp. 7313– 7331, 2021. [27] M. Ashraf, G. Geng, X. Wang, F. Ahmad, and F. Abid, “A globally regu- larized joint neural architecture for music classification,” IEEE Access, vol. 8, pp. 220980–220989, 2020. [28] M. Agrawal and A. Nandy, “A novel multimodal music genre classi- fier using hierarchical attention and convolutional neural network,” ArXiv, vol. abs/2011.11970, 2020. [29] S. Allamy and A. L. Koerich, “1d cnn architectures for music genre classifica- tion,” ArXiv, vol. abs/2105.07302, 2021. [30] Q. Li, H. Peng, J. Li, C. Xia, R. Yang, L. Sun, P. S. Yu, and L. He, “A survey on text classification: From shallow to deep learning,” ArXiv, vol. abs/2008.00364, 2020. [31] P. Zhou, Z. Qi, S. Zheng, J. Xu, H. Bao, and B. Xu, “Text classification im- proved by integrating bidirectional lstm with two-dimensional max pooling,” in COLING, 2016. [32] P. Liu, X. Qiu, and X. Huang, “Recurrent neural network for text classifi- cation with multi-task learning,” in Proceedings of the Twenty-Fifth Inter- national Joint Conference on Artificial Intelligence, IJCAI’16, p. 2873–2879, AAAI Press, 2016. [33] D. Wang, J. Gong, and Y. xi Song, “W-rnn: News text classification based on a weighted rnn,” ArXiv, vol. abs/1909.13077, 2019. [34] D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by jointly learning to align and translate,” Jan. 2015. 3rd International Conference on Learning Representations, ICLR 2015 ; Conference date: 07-05-2015 Through 09-05-2015. [35] R. Paulus, C. Xiong, and R. Socher, “A deep reinforced model for abstrac- tive summarization,” in International Conference on Learning Representations, 2018. [36] J. Gehring, M. Auli, D. Grangier, D. Yarats, and Y. Dauphin, “Convolutional sequence to sequence learning,” in ICML, 2017. [37] S. Woo, J. Park, J.-Y. Lee, and I.-S. Kweon, “Cbam: Convolutional block attention module,” in ECCV, 2018. [38] L. Fenaux and M. J. Quintero, “Bumblebee: A transformer for music,” ArXiv, vol. abs/2107.03443, 2021. [39] Z. Dai, Z. Yang, Y. Yang, J. G. Carbonell, Q. V. Le, and R. Salakhutdinov, “Transformer-xl: Attentive language models beyond a fixed-length context,” in ACL, 2019. [40] J. Libovický, J. Helcl, and D. Mareček, “Input combination strategies for multi- source transformer decoder,” in WMT, 2018. [41] Z. Wang, K. Chen, J. Jiang, Y. Zhang, M. Xu, S. Dai, X. Gu, and G. Xia, “Pop909: A pop-song dataset for music arrangement generation,” in ISMIR, 2020. [42] L.-C. Yang and A. Lerch, “On the evaluation of generative models in music,” Neural Computing and Applications, vol. 32, no. 9, pp. 4773–4784, 2020. [43] R. Tavenard, J. Faouzi, G. Vandewiele, F. Divo, G. Androz, C. Holtz, M. Payne, R. Yurchak, M. Rußwurm, K. Kolar, and E. Woods, “Tslearn, a machine learn- ing toolkit for time series data,” Journal of Machine Learning Research, vol. 21, no. 118, pp. 1–6, 2020. [44] B. Paassen, B. Mokbel, and B. Hammer, “A toolbox for adaptive sequence dissimilarity measures for intelligent tutoring systems,” in EDM, 2015. [45] V. I. Levenshtein, “Binary Codes Capable of Correcting Deletions, Insertions and Reversals,” Soviet Physics Doklady, vol. 10, p. 707, Feb. 1966. [46] H. Sakoe and S. Chiba, “Dynamic programming algorithm optimization for spoken word recognition,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 26, no. 1, pp. 43–49, 1978. [47] R. Popovici and R. Andonie, “Music genre classification with self-organizing maps and edit distance,” in 2015 International Joint Conference on Neural Networks (IJCNN), pp. 1–7, 2015. [48] B. Benward and M. N. Saker, Music in theory and practice. McGraw-Hill Education, 2015. [49] C. Donahue, H. H. Mao, Y. Li, G. Cottrell, and J. McAuley, “Lakhnes: Im- proving multi-instrumental music generation with cross-domain pre-training,” in ISMIR, 2019. [50] M. S. Cuthbert and C. Ariza, “Music21: A toolkit for computer-aided musicol- ogy and symbolic music data.,” in ISMIR (J. S. Downie and R. C. Veltkamp, eds.), pp. 637–642, International Society for Music Information Retrieval, 2010. [51] O. M. Bjørndalen, “Midi objects for python¶.” https://mido.readthedocs. io/en/latest; accessed 5-Dec-2021. [52] M. Schwenk, “Midieditor: Graphical interface to edit, play, and record midi data.” https://midieditor.org/; accessed 5-Dec-2021. |