帳號:guest(216.73.216.146)          離開系統
字體大小: 字級放大   字級縮小   預設字形  

詳目顯示

以作者查詢圖書館館藏以作者查詢臺灣博碩士論文系統以作者查詢全國書目
作者(中文):熊磊
作者(外文):Hsiung, Lei
論文名稱(中文):對抗複合擾動之穩健深度學習
論文名稱(外文):Robust Deep Learning Against Composite Perturbations
指導教授(中文):何宗易
指導教授(外文):Ho, Tsung-Yi
口試委員(中文):游家牧
陳品諭
口試委員(外文):Yu, Chia-Mu
Chen, Pin-Yu
學位類別:碩士
校院名稱:國立清華大學
系所名稱:資訊工程學系
學號:109062509
出版年(民國):111
畢業學年度:110
語文別:英文
論文頁數:63
中文關鍵詞:深度學習複合對抗攻擊複合擾動泛化對抗訓練穩健性評估穩健性基準
外文關鍵詞:Deep LearningComposite Adversarial AttackComposite PerturbationsGeneralized Adversarial TrainingRobustness EvaluationRobustness Benchmark
相關次數:
  • 推薦推薦:0
  • 點閱點閱:0
  • 評分評分:*****
  • 下載下載:0
  • 收藏收藏:0
過去關於對抗攻擊的研究主要集中在單一威脅的攻擊與防禦方式(例如 ℓp 空間下的有界擾動),而將對抗攻擊推廣到接近真實情況的語義擾動(例如色調、飽和度、亮度、對比度和旋轉),或多個威脅的複合擾動時,仍存在許多研究空間。本論文首先提出了生成複合對抗樣本的新方法-複合對抗攻擊。除了擴展圖像的可擾空間,複合對抗攻擊可以利用組件級梯度下降法及攻擊順序自動排程,找到最佳的攻擊組合。本論文進而提出泛化對抗訓練,將模型穩健性從 ℓp 空間擴展到複合語義空間。在 ImageNet 和 CIFAR-10 資料集上的實驗結果顯示,泛化對抗訓練不僅可以抵抗單一攻擊,亦可抵抗來自多種攻擊的複合擾動;同時,其抗擾性甚優於僅在 ℓ∞ 空間內進行對抗訓練之穩健模型。另一方面,複合對抗攻擊的實驗顯示當前評估穩健模型的方式未臻完善。因此,本論文研發網路應用程式-CARBEN,展示在複合對抗攻擊中,攻擊順序及其擾動程度如何影響結果圖像,並提供互動介面與各式模型的實時預測,幫助研究者快速評估神經網路模型的表現。本論文最後提出針對複合對抗攻擊,穩健模型的排行榜,可與其它對抗訓練方法進行比較,以更全面地評估模型穩健性。
Prior literature on adversarial attack methods has mainly focused on attacking with and defending against a single threat model, e.g., perturbations bounded in ℓp-ball. Thus, more realistic scenarios involving multiple semantic perturbations (e.g., perturbations of hue, saturation, brightness, contrast, and rotation) and their compositions have hitherto remained largely unexplored. To address that gap, this thesis firstly proposes a novel method for generating composite adversarial examples, composite adversarial attack (CAA), which expands the perturbable space of the image and identifies the optimal attack composition by utilizing component-wise projected gradient descent and automatic attack-order scheduling. Next, it proposes generalized adversarial training (GAT) to extend model robustness from ℓp-norm to composite semantic perturbations. Results obtained using ImageNet and CIFAR-10 datasets indicate that GAT can be robust not only to all the tested types of single attack, but also to any combination of such attacks. GAT also outperformed baseline ℓ∞-norm bounded adversarial training approaches by a significant margin. On the other hand, CAA may be overlooked by current modes of robustness evaluation. Third, this thesis proposes a web application, composite adversarial robustness benchmark (CARBEN). As well as demonstrating how CAA's attack order and perturbation level affect the resulting image, CARBEN provides interactive interfaces and real-time inferences about different models, to facilitate researchers' rapid evaluation of model prediction. Lastly, a leaderboard to benchmark adversarial robustness against CAA is also introduced to compete with other state-of-the-art adversarial training approaches for a more comprehensive assessment of model robustness.
Abstract
Contents
Chapter 1 Introduction --- 1
Chapter 2 Related Work --- 5
Chapter 3 Methodologies --- 9
Chapter 4 Experiments --- 19
Chapter 5 CARBEN: Composite Adversarial Robustness Benchmark --- 31
Chapter 6 Conclusions --- 41
Chapter 7 Appendix --- 43
Bibliography --- 57
[1] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778, 2016.
[2] A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu, “Towards deep learn-ing models resistant to adversarial attacks,” International Conference on Learning Representations, 2018.
[3] H. Zhang, Y. Yu, J. Jiao, E. P. Xing, L. E. Ghaoui, and M. I. Jordan, “Theoretically principled trade-off between robustness and accuracy,” in International Conference on Machine Learning, 2019.
[4] S. Zagoruyko and N. Komodakis, “Wide residual networks,” in Proceedings of the British Machine Vision Conference (BMVC) (E. R. H. Richard C. Wilson and W. A. P. Smith, eds.), pp. 87.1–87.12, BMVA Press, September 2016.
[5] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus, “Intriguing properties of neural networks,” International Conference on Learning Representations, 2014.
[6] C. Laidlaw, S. Singla, and S. Feizi, “Perceptual adversarial robustness: Defense against unseen threat models,” in ICLR, 2021.
[7] J. Zhang, X. Xu, B. Han, G. Niu, L. Cui, M. Sugiyama, and M. Kankanhalli, “Attacks which do not kill training make adversarial learning stronger,” in International Conference on Machine Learning, 2020.
[8] D. Wu, S.-T. Xia, and Y. Wang, “Adversarial weight perturbation helps robust generalization,” in Advances in Neural Information Processing Systems, 2020.
[9] E. Wong, L. Rice, and J. Z. Kolter, “Fast is better than free: Revisiting adversarial training,” in International Conference on Learning Representations, 2020.
[10] F. Croce, M. Andriushchenko, V. Sehwag, E. Debenedetti, N. Flammarion, M. Chi-ang, P. Mittal, and M. Hein, “Robustbench: a standardized adversarial robustness benchmark,” in Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2021.
[11] B. Biggio and F. Roli, “Wild patterns: Ten years after the rise of adversarial machine learning,” Pattern Recognit., vol. 84, pp. 317–331, 2018.
[12] I. J. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” International Conference on Learning Representations, 2015.
[13] N. Carlini and D. A. Wagner, “Towards evaluating the robustness of neural net-works,” in 2017 IEEE Symposium on Security and Privacy, SP 2017, San Jose, CA, USA, May 22-26, 2017, pp. 39–57, IEEE Computer Society, 2017.
[14] P.-Y. Chen, Y. Sharma, H. Zhang, J. Yi, and C.-J. Hsieh, “EAD: elastic-net attacks to deep neural networks via adversarial examples,” in Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, AAAI’18/IAAI’18/EAAI’18, AAAI Press, 2018.
[15] F. Croce and M. Hein, “Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks,” in International conference on machine learning, pp. 2206–2216, PMLR, 2020.
[16] H. Hosseini and R. Poovendran, “Semantic adversarial examples,” in IEEE Con-ference on Computer Vision and Pattern Recognition Workshops, pp. 1614–1619, 2018.
[17] A. Joshi, A. Mukherjee, S. Sarkar, and C. Hegde, “Semantic adversarial attacks: Para-metric transformations that fool deep classifiers,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4773–4783, 2019.
[18] A. S. Shamsabadi, R. Sanchez-Matilla, and A. Cavallaro, “Colorfool: Semantic adversarial colorization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1151–1160, 2020.
[19] Y. Wang, S. Wu, W. Jiang, S. Hao, Y.-a. Tan, and Q. Zhang, “Demiguise attack: Crafting invisible semantic adversarial perturbations with perceptual similarity,” arXiv preprint arXiv:2107.01396, 2021.
[20] S. Wang, S. Chen, T. Chen, S. Nepal, C. Rudolph, and M. Grobler, “Gener-ating semantic adversarial examples via feature manipulation,” arXiv preprint arXiv:2001.02297, 2020.
[21] A. Bhattad, M. J. Chong, K. Liang, B. Li, and D. A. Forsyth, “Unrestricted ad-versarial examples via semantic manipulation,” in 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020, OpenReview.net, 2020.
[22] D. Kang, Y. Sun, D. Hendrycks, T. Brown, and J. Steinhardt, “Testing robustness against unforeseen adversaries,” arXiv preprint arXiv:1908.08016, 2019.
[23] H. Qiu, C. Xiao, L. Yang, X. Yan, H. Lee, and B. Li, “Semanticadv: Generating adver-sarial examples via attribute-conditioned image editing,” in European Conference on Computer Vision, pp. 19–37, Springer, 2020.
[24] C. Xiao, J.-Y. Zhu, B. Li, W. He, M. Liu, and D. Song, “Spatially transformed adversarial examples,” arXiv preprint arXiv:1801.02612, 2018.
[25] L. Engstrom, B. Tran, D. Tsipras, L. Schmidt, and A. Madry, “Exploring the land-scape of spatial robustness,” in International Conference on Machine Learning, pp. 1802–1811, PMLR, 2019.
[26] E. Wong, F. Schmidt, and Z. Kolter, “Wasserstein adversarial examples via projected sinkhorn iterations,” in International Conference on Machine Learning, pp. 6808–6817, PMLR, 2019.
[27] J. Mohapatra, T.-W. Weng, P.-Y. Chen, S. Liu, and L. Daniel, “Towards verifying robustness of neural networks against a family of semantic perturbations,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
[28] I. Dunn, L. Hanu, H. Pouget, D. Kroening, and T. Melham, “Evaluating robustness to context-sensitive feature perturbations of different granularities,” arXiv preprint arXiv:2001.11055, 2020.
[29] D. Zhou, T. Liu, B. Han, N. Wang, C. Peng, and X. Gao, “Towards defending against adversarial examples via attack-invariant features,” ArXiv, vol. abs/2106.05036, 2021.
[30] C. Laidlaw and S. Feizi, “Functional adversarial attacks,” arXiv preprint arXiv:1906.00001, 2019.
[31] M. Jordan, N. Manoj, S. Goel, and A. G. Dimakis, “Quantifying perceptual distortion of adversarial examples,” arXiv preprint arXiv:1902.08265, 2019.
[32] X. Mao, Y. Chen, S. Wang, H. Su, Y. He, and H. Xue, “Composite adversarial attacks,” Association for the Advancement of Artificial Intelligence (AAAI), 2021.
[33] A. Kurakin, I. Goodfellow, and S. Bengio, “Adversarial machine learning at scale,” International Conference on Learning Representations, 2017.
[34] D. Stutz, M. Hein, and B. Schiele, “Confidence-calibrated adversarial training: Generalizing to unseen attacks,” in International Conference on Machine Learning, pp. 9155–9166, PMLR, 2020.
[35] Y. Sharma and P.-Y. Chen, “Attacking the Madry defense model with L1-based adversarial examples,” ICLR Workshop, 2018.
[36] F. Tramèr and D. Boneh, “Adversarial training and robustness for multiple pertur-bations,” arXiv preprint arXiv:1904.13000, 2019.
[37] J. Wang, T. Zhang, S. Liu, P.-Y. Chen, J. Xu, M. Fardad, and B. Li, “Towards a unified min-max framework for adversarial exploration and robustness,” arXiv preprint arXiv:1906.03563, 2019.
[38] P. Maini, E. Wong, and Z. Kolter, “Adversarial robustness against the union of multiple perturbation models,” in Proceedings of the 37th International Conference on Machine Learning (H. D. III and A. Singh, eds.), vol. 119 of Proceedings of Machine Learning Research, pp. 6640–6650, PMLR, 13–18 Jul 2020.
[39] G. Mena, D. Belanger, S. Linderman, and J. Snoek, “Learning latent permuta-tions with gumbel-sinkhorn networks,” in International Conference on Learning Representations, 2018.
[40] R. Sinkhorn, “A relationship between arbitrary positive matrices and stochastic matrices,” Canadian Journal of Mathematics, vol. 18, p. 303–306, 1966.
[41] H. Wang and A. Banerjee, “Bregman alternating direction method of multipliers,” in Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, NeurIPS’14, (Cambridge, MA, USA), p. 2816–2824, MIT Press, 2014.
[42] R. Sinkhorn and P. Knopp, “Concerning nonnegative matrices and doubly stochas-tic matrices,” Pacific Journal of Mathematics, vol. 21, no. 2, pp. 343–348, 1967.
[43] H. W. Kuhn, “The hungarian method for the assignment problem,” Naval research logistics quarterly, vol. 2, no. 1-2, pp. 83–97, 1955.
[44] J. R. Munkres, “Algorithms for the assignment and transportation problems,” Jour-nal of The Society for Industrial and Applied Mathematics, vol. 10, pp. 196–210, 1957.
[45] M. Cuturi, “Sinkhorn distances: Lightspeed computation of optimal transport,” Advances in neural information processing systems, vol. 26, pp. 2292–2300, 2013.
[46] J. Altschuler, J. Weed, and P. Rigollet, “Near-linear time approximation algorithms for optimal transport via sinkhorn iteration,” in Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp. 1964–1974, 2017.
[47] A. Kurakin, I. Goodfellow, and S. Bengio, “Adversarial examples in the physical world,” ICLR Workshop, 2017.
[48] E. Riba, D. Mishkin, D. Ponsa, E. Rublee, and G. Bradski, “Kornia: an open source differentiable computer vision library for pytorch,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3674–3683, 2020.
[49] A. Krizhevsky and G. Hinton, “Learning multiple layers of features from tiny images,” Tech. Rep. 0, University of Toronto, Toronto, Ontario, 2009.
[50] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “ImageNet Large Scale Visual Recognition Challenge,” International Journal of Computer Vision (IJCV), vol. 115, no. 3, pp. 211–252, 2015.
[51] L. Engstrom, A. Ilyas, H. Salman, S. Santurkar, and D. Tsipras, “Robustness (python library),” 2019.
[52] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Kopf, E. Yang, Z. DeVito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala, “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems 32 (H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, eds.), pp. 8024–8035, Curran Associates, Inc., 2019.
(此全文未開放授權)
電子全文
摘要
 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
* *