|
[1] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” in Proceedings of International Conference on Learning Representations, 2021. [2] D. Parikh, “Recognizing jumbled images: The role of local and global information in image classification,” in Proceedings of International Conference on Computer Vision, pp. 519–526, 2011. [3] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. U. Kaiser, and I. Polosukhin, “Attention is all you need,” in Proceedings of Neural Information Processing Systems, pp. 5999–6009, 2017. [4] M. Nagel, M. Fournarakis, R. A. Amjad, Y. Bondarenko, M. Van Baalen, and T. Blankevoort, “A white paper on neural network quantization,” arXiv preprint arXiv:2106.08295, 2021. [5] M. H. Ahmadilivani, M. Taheri, J. Raik, M. Daneshtalab, and M. Jenihhin, “A systematic literature review on hardware reliability assessment methods for deep neural networks,” ACM Computing Surveys, vol. 56, no. 6, pp. 141:1–141:39, 2024. [6] A. Ruospo, L. M. Luza, A. Bosio, M. Traiola, L. Dilillo, and E. Sanchez, “Pros and cons of fault injection approaches for the reliability assessment of deep neural networks,” in Proceedings of IEEE Latin American Test Symposium, pp. 87–91, 2021. [7] G. Bolt, “Fault models for artificial neural networks,” in Proceedings of IEEE International Joint Conference on Neural Networks, pp. 1373–1378, 1991. [8] C. Liu, Z. Gao, S. Liu, X. Ning, H. Li, and X. Li, “Fault-tolerant deep learning: A hierarchical perspective,” arXiv preprint arXiv:2204.01942, 2022. [9] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255, 2009. [10] H. Touvron, M. Cord, M. Douze, F. Massa, A. Sablayrolles, and H. Jegou, “Training data-efficient image transformers & distillation through attention,” in Proceedings of International Conference on Machine Learning, pp. 10347–10357, 2021. [11] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of International Conference on Computer Vision, pp. 9992–10002, 2021. [12] S. Xie, R. B. Girshick, P. Dollár, Z. Tu, and K. He, “Aggregated residual transformations for deep neural networks,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 5987–5995, 2017. [13] M. Hasan and B. Ray, “Tolerance of deep neural network against the bit error rate of nand flash memory,” in Proceedings of IEEE International Reliability Physics Symposium, pp. 814–817, 2019. [14] M. M. Hasan and B. Ray, “Reliability of nand flash memory as a weight storage device of artificial neural network,” IEEE Transactions on Device and Materials Reliability, vol. 20, no. 3, pp. 596–603, 2020. [15] B. Sangchoolie, K. Pattabiraman, and J. Karlsson, “One bit is (not) enough: An empirical study of the impact of single and multiple bit-flip errors,” in Proceedings of IEEE International Conference on Dependable Systems and Networks, pp. 97–108, 2017. [16] Z. Chen, G. Li, K. Pattabiraman, and N. DeBardeleben, “Binfi: An efficient fault injector for safety-critical machine learning systems,” in Proceedings of International Conference for High Performance Computing, Networking, Storage and Analysis, pp. 69:1–69:23, 2019. [17] Z. Chen, G. Li, and K. Pattabiraman, “A low-cost fault corrector for deep neural networks through range restriction,” in Proceedings of IEEE International Conference on Dependable Systems and Networks, pp. 1–13, 2021. [18] Y. Lin, T. Zhang, P. Sun, Z. Li, and S. Zhou, “Fq-vit: Post-training quantization for fully quantized vision transformer,” in Proceedings of International Joint Conference on Artificial Intelligence, pp. 1173–1179, 2022. [19] I. S. Haque and V. S. Pande, “Hard data on soft errors: A large-scale assessment of real-world error rates in gpgpu,” in Proceedings of IEEE International Conference on Cluster, Cloud and Grid Computing, pp. 691–696, 2010. |