|
[1] Y. Lei, N. Scheffer, L. Ferrer, and M. McLaren, “A novel scheme for speaker recognition using a phonetically-aware deep neural network,” in Proc. International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1695–1699, 2014. [2] N. Carlini and D. A. Wagner, “Audio adversarial examples: targeted attacks on speech-to-text,” CoRR, vol. abs/1801.01944, 2018. [3] C. Anirban, A. Manaar, D. Vishal, C. Anupam, and M. Debdeep, “Adversarial attacks and defences: A survey,” CoRR, vol. abs/1810.00069, 2018. [4] Y. Zhang, N. Suda, L. Lai, and V. Chandra, “Hello edge: keyword spotting on microcontrollers,” CoRR, vol. abs/1711.07128, 2017. [5] C. Teacher, H. Kellett, and L. Focht, “Experimental, limited vocabulary, speech recognizer,” IEEE Transactions on Audio and Electroacoustics, vol. 15, pp. 127–130, 1967. [6] J. Wilpon, L. Rabiner, C.-H. Lee, and E. Goldman, “Automatic recognition of keywords in unconstrained speech using hidden markov models,” IEEE Transactions on Audio and Electroacoustics, vol. 38, pp. 1870–1878, 1990. 36 [7] R. C. Rose and D. B. Paul, “A hidden markov model based keyword recognition system,” in Proc. International Conference on Acoustics, Speech, and Signal Processing (ICASSP), pp. 129–132, 1990. [8] S. Fern´andez, A. Graves, and J. Schmidhuber, “An application of recurrent neural networks to discriminative keyword spotting,” in Proc. Internet Corporation for Assigned Names and Numbers (ICANN), pp. 220–229, 2007. [9] G. Tucker, M. Wu, M. Sun, S. Panchapagesan, G. Fu, and S. Vitaladevuni, “Model compression applied to small-footprint keyword spotting,” in Proc. International Speech Communication Association (INTERSPEECH), pp. 1878–1882, 2016. [10] G. Nakkiran, R. Alvarez, R. Prabhavalkar, and C. Parada, “Compressing deep neural networks using a rank-constrained topology,” in Proc. International Speech Communication Association (INTERSPEECH), pp. 1473—-1477, 2015. [11] N. Carlini and D. A. Wagner, “Towards evaluating the robustness of neural networks,” in Proc. IEEE Symposium on Security and Privacy (SP), pp. 39–57, 2017. [12] A. Hannun, C. Case, J. Casper, B. Catanzaro, G. Diamos, E. Elsen, R. Prenger, S. Satheesh, S. Sengupta, A. Coates, and A. Y. Ng, “Deep speech: Scaling up end-to-end speech recognition,” CoRR, vol. abs/1412.5567, 2014. [13] Y. Qin, N. Carlini, I. Goodfellow, G. Cottrell, and C. Raffel, “Imperceptible, robust, and targeted adversarial examples for automatic speech recognition,” in Proc. International Conference on Machine Learning (ICML), 2019. 37 [14] V. Subramanian, E. Benetos, N. Xu, S. McDonald, and M. Sandler, “Adversarial attacks in sound event classification,” CoRR, vol. abs/1907.02477, 2019. [15] M. Alzantot, B. Balaji, and M. B. Srivastava, “Did you hear that? adversarial examples against automatic speech recognition,” CoRR, vol. abs/1801.00554, 2018. [16] R. Taori, A. Kamsetty, B. Chu, and N. Vemuri, “Targeted adversarial examples for black box audio systems,” CoRR, vol. abs/1805.07820, 2018. [17] J. Vadillo and R. Santana, “Universal adversarial examples in speech command classification,” CoRR, vol. abs/1911.10182, 2019. [18] K.-H. Chang, P.-H. Huang, H. Yu, Y. Jin, and T.-C. Wang, “Audio adversarial examples generation with recurrent neural networks,” in Proc. Asia and South Pacific Design Automation Conference (ASP-DAC), pp. 488–493, 2020. [19] Y. Gong, B. Li, C. Poellabauer, and Y. Shi, “Real-time adversarial attacks,” in Proc. International Joint Conference on Artificial Intelligence (IJCAI), pp. 4672–4680, 2019. [20] Z. Yang, B. Li, P.-Y. Chen, and D. Song, “Towards mitigating audio adversarial perturbations,” in Proc. International Conference on Learning Representations (ICLR) workshop, 2018. [21] K. Rajaratnam, K. Shah, and J. Kalita, “Isolated and ensemble audio preprocessing methods for detecting adversarial examples against automatic speech recognition,” in Proc. Conference on Computational Linguistics and Speech Processing (ROCLING), 2018. 38 [22] I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” CoRR, vol. abs/1406.2661, 2014. [23] P.-Y. Chen, H. Zhang, Y. Sharma, J. Yi, and C.-J. Hsieh, “Zoo: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models,” in Proc. ACM Workshop on Artificial Intelligence and Security, pp. 15–26, 2017. [24] C.-C. Tu, P. Ting, P.-Y. Chen, S. Liu, H. Zhang, J. Yi, C.-J. Hsieh, and S.-M. Cheng, “Autozoom: Autoencoder-based zeroth order optimization method for attacking black-box neural networks,” in Proc. AAAI, 2019. [25] F. Tramer, N. Carlini, W. Brendel, and A. Madry, “On adaptive attacks to adversarial example defenses,” CoRR, vol. abs/2002.08347, 2020. [26] J. Yang, Q. Zhang, R. Fang, B. Ni, J. Liu, and Q. Tian, “Adversarial attack and defense on point sets,” CoRR, vol. abs/1902.10899, 2019. [27] K. Eykholt, I. Evtimov, E. Fernandes, B. Li, A. Rahmati, F. Tramer, A. Prakash, T. Kohno, and D. Song, “Physical adversarial examples for object detectors,” CoRR, vol. abs/1807.07769, 2018. [28] “Speech commands dataset." https://research.googleblog.com/2017/08/launchingspeech-commands-dataset.html. [29] H. Yakura and J. Sakuma, “Robust audio adversarial example for a physical attack,” CoRR, vol. abs/1810.11793, 2018. [30] I. J. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” CoRR, vol. abs/1412.6572, 2015. |