|
[1] Fabian-Robert St ̈oter, Antoine Liutkus, and Nobutaka Ito. The 2018 signal separationevaluation campaign.ArXiv, abs/1804.06267, 2018. [2] Emmanuel Vincent, R ́emi Gribonval, and C ́edric F ́evotte. Performance measurementin blind audio source separation.IEEE Transactions on Audio, Speech, and LanguageProcessing, 14:1462–1469, 2006. [3] Zafar Rafii, Antoine Liutkus, Fabian-Robert St ̈oter, Stylianos Ioannis Mimilakis, andRachel M. Bittner. Musdb18 - a corpus for music separation. 2017. [4] Pritish Chandna, Merlijn Blaauw, Jordi Bonada, and Emilia G ́omez. A vocoder basedmethod for singing voice extraction.ICASSP 2019 - 2019 IEEE International Con-ference on Acoustics, Speech and Signal Processing (ICASSP), pages 990–994, 2019. [5] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networksfor biomedical image segmentation.ArXiv, abs/1505.04597, 2015. [6] Andreas Jansson, Eric J. Humphrey, Nicola Montecchio, Rachel M. Bittner, AparnaKumar, and Tillman Weyde. Singing voice separation with deep u-net convolutionalnetworks. InISMIR, 2017. [7] Naoya Takahashi, Nabarun Goswami, and Yuki Mitsufuji. Mmdenselstm: An effi-cient combination of convolutional and recurrent neural networks for audio sourceseparation.2018 16th International Workshop on Acoustic Signal Enhancement(IWAENC), pages 106–110, 2018. [8] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networksfor semantic segmentation. InCVPR, 2015. [9] Zhengxin Zhang, Qingjie Liu, and Yunhong Wang. Road extraction by deep residualu-net.IEEE Geoscience and Remote Sensing Letters, 15:749–753, 2018. [10] ̈Ozg ̈un iek, Ahmed Abdulkadir, Soeren S. Lienkamp, Thomas Brox, and Olaf Ron-neberger. 3d u-net: Learning dense volumetric segmentation from sparse annotation.InMICCAI, 2016. [11] Stefan Uhlich, Marcello Porcu, Franck Giron, Michael Enenkl, Thomas Kemp, NaoyaTakahashi, and Yuki Mitsufuji. Improving music source separation based on deepneural networks through data augmentation and network blending.2017 IEEE In-ternational Conference on Acoustics, Speech and Signal Processing (ICASSP), pages261–265, 2017. [12] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott E. Reed, DragomirAnguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Goingdeeper with convolutions.2015 IEEE Conference on Computer Vision and PatternRecognition (CVPR), pages 1–9, 2014. [13] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learn-ing for image recognition.2016 IEEE Conference on Computer Vision and PatternRecognition (CVPR), pages 770–778, 2015. [14] A ̈aron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals,Alex Graves, Nal Kalchbrenner, Andrew W. Senior, and Koray Kavukcuoglu.Wavenet: A generative model for raw audio. InSSW, 2016. |