|
[1] Ga ̈etan Hadjeres, F. Pachet, and F. Nielsen. Deepbach: a steerable model for bachchorales generation. InICML, 2017. [2] Liangrong Yi and J. Goldsmith. Automatic generation of four-part harmony. InBMA,2007. [3] J. Yosinski, J. Clune, Yoshua Bengio, and Hod Lipson. How transferable are featuresin deep neural networks?ArXiv, abs/1411.1792, 2014. [4] Sageev Oore, I. Simon, S. Dieleman, D. Eck, and K. Simonyan. This time with feel-ing: learning expressive musical performance.Neural Computing and Applications,pages 1–13, 2018. [5] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan NGomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In I. Guyon,U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Gar-nett, editors,Advances in Neural Information Processing Systems, volume 30. CurranAssociates, Inc., 2017. URLhttps://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf [6] Cheng-Zhi Anna Huang, Ashish Vaswani, Jakob Uszkoreit, I. Simon, CurtisHawthorne, Noam M. Shazeer, Andrew M. Dai, M. Hoffman, M. Dinculescu, andD. Eck. Music transformer: Generating music with long-term structure. InICLR,2019. [7] Zihang Dai, Z. Yang, Yiming Yang, J. Carbonell, Quoc V. Le, and R. Salakhutdinov.Transformer-xl: Attentive language models beyond a fixed-length context. InACL,2019. [8] Peter Shaw, Jakob Uszkoreit, and Ashish Vaswani. Self-attention with relative posi-tion representations. InNAACL-HLT, 2018. [9] Zachary M. Ziegler, Luke Melas-Kyriazi, Sebastian Gehrmann, and Alexander M.Rush. Encoder-agnostic adaptation for conditional language generation.ArXiv,abs/1908.06938, 2019. [10] I. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley,S. Ozair, Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. InNIPS, 2014. [11] Diederik P. Kingma and M. Welling.Auto-encoding variational bayes.CoRR,abs/1312.6114, 2014. [12] S. Hochreiter and J. Schmidhuber. Long short-term memory.Neural Computation, 9:1735–1780, 1997. [13] J. Chung, aglar G ̈ulehre, Kyunghyun Cho, and Yoshua Bengio. Empirical evaluation of gated recurrent neural networks on sequence modeling.ArXiv, abs/1412.3555,2014. [14] ElliotWaite.Generatinglong-termstructureinsongsandsto-ries.https://magenta.tensorflow.org/2016/07/15/lookback-rnn-attention-rnn. 2016. [15] D. Johnson. Generating polyphonic music using tied parallel networks. InEvo-MUSART, 2017. [16] Li-Chia Yang, Szu-Yu Chou, and Y. Yang. Midinet: A convolutional generative ad-versarial network for symbolic-domain music generation. InISMIR, 2017. [17] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learn-ing with deep convolutional generative adversarial networks.CoRR, abs/1511.06434,2016. [18] Hao-Wen Dong, Wen-Yi Hsiao, Li-Chia Yang, and Y. Yang. Musegan: Multi-tracksequential generative adversarial networks for symbolic music generation and accom-paniment. In AAAI, 2018. [19] Ishaan Gulrajani, Faruk Ahmed, Mart ́ın Arjovsky, Vincent Dumoulin, and Aaron C.Courville. Improved training of wasserstein gans. InNIPS, 2017. [20] Geoffrey E. Hinton and R. Salakhutdinov. Reducing the dimensionality of data withneural networks.Science, 313:504 – 507, 2006. [21] Adam Roberts, Jesse Engel, Colin Raffel, Curtis Hawthorne, and D. Eck. A hi-erarchical latent vector model for learning long-term structure in music.ArXiv,abs/1803.05428, 2018. [22] Yu-Siang Huang and Y. Yang. Pop music transformer: Generating music with rhythmand harmony.ArXiv, abs/2002.00212, 2020. [23] H. H. Mao, Taylor Shin, and G. Cottrell. Deepj: Style-specific music generation.2018IEEE 12th International Conference on Semantic Computing (ICSC), pages 377–382,2018. [24] Gino Brunner, Andres Konrad, Y. Wang, and Roger Wattenhofer. Midi-vae: Modelingdynamics and instrumentation of music with applications to style transfer.ArXiv,abs/1809.07600, 2018. [25] Yu-Quan Lim, Chee Seng Chan, and F. Loo. Clavinet: Generate music with differentmusical styles.IEEE MultiMedia, 28:83–93, 2021. [26] Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-softmax.ArXiv, abs/1611.01144, 2017. [27] J. Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-trainingof deep bidirectional transformers for language understanding. InNAACL-HLT, 2019. [28] Alec Radford, Jeffrey Wu, R. Child, David Luan, Dario Amodei, and Ilya Sutskever.Language models are unsupervised multitask learners. 2019. [29] T. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, J. Kaplan, Prafulla Dhari-wal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, T. Henighan, R. Child, A. Ramesh,Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, EricSigler, Mateusz Litwin, Scott Gray, Benjamin Chess, J. Clark, Christopher Berner,Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language mod-els are few-shot learners.ArXiv, abs/2005.14165, 2020. [30] J. Russell. A circumplex model of affect.Journal of Personality and Social Psychol-ogy, 39:1161–1178, 1980. [31] Li-Chia Yang and Alexander Lerch. On the evaluation of generative models in music.Neural Computing and Applications, 32:4773–4784, 2018. [32] Shih-Lun Wu and Yi-Hsuan Yang. The jazz transformer on the front line: Explor-ing the shortcomings of ai-composed music through quantitative measures.ArXiv,abs/2008.01307, 2020. [33] Hao-Wen Dong, K. Chen, Julian McAuley, and Taylor Berg-Kirkpatrick. Muspy: Atoolkit for symbolic music generation.ArXiv, abs/2008.01951, 2020.[34] Curtis Hawthorne, Andriy Stasyuk, Adam Roberts, Ian Simon, Cheng-Zhi AnnaHuang, Sander Dieleman, Erich Elsen, Jesse Engel, and Douglas Eck. Enablingfactorized piano music modeling and generation with the MAESTRO dataset. InInternational Conference on Learning Representations, 2019.URLhttps://openreview.net/forum?id=r1lYRjC9F7. [35] Lucas N Ferreira, Levi HS Lelis, and Jim Whitehead. Computer-generated music fortabletop role-playing games. 2020. [36] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization.CoRR, abs/1412.6980, 2015. |