帳號:guest(216.73.216.146)          離開系統
字體大小: 字級放大   字級縮小   預設字形  

詳目顯示

以作者查詢圖書館館藏以作者查詢臺灣博碩士論文系統以作者查詢全國書目
作者(中文):黃詮舜
作者(外文):Huang, Chuan-Shun
論文名稱(中文):一個有效率的迭代式網路剪枝方法
論文名稱(外文):An Efficient Approach to Iterative Network Pruning
指導教授(中文):王俊堯
指導教授(外文):Wang, Chun-Yao
口試委員(中文):黃俊達
陳宏明
口試委員(外文):Huang, Juinn-Dar
Chen, Hung-Ming
學位類別:碩士
校院名稱:國立清華大學
系所名稱:資訊工程學系
學號:110062553
出版年(民國):112
畢業學年度:111
語文別:英文
論文頁數:19
中文關鍵詞:神經網絡剪枝
外文關鍵詞:Neural Network Pruning
相關次數:
  • 推薦推薦:0
  • 點閱點閱:198
  • 評分評分:*****
  • 下載下載:0
  • 收藏收藏:0
神經網絡剪枝是一種使大型神經網絡之參數量減少的技術。神經網絡剪枝可以分為一次性或迭代式進行。一次性的網絡剪枝可以快速得到所需的稀疏度,但在追求高稀疏度的情況下可能會出現無可挽回的準確度下降。另一方面,迭代式網絡剪枝以多次的少量網路修剪搭配重新訓練網絡以維持準確度,但這種重覆性流程所需要的運算時間非常長。在這篇論文中,我們提出了一種高效的神經網絡剪枝方法。實驗結果顯示,我們的方法與最先進的迭代式網絡剪枝方法相比,減少了25%至近60%的運算時間且得到相近的網路準確度。
Network pruning is a technique to minimize the number of parameters of large neural networks. Network pruning can be performed once or multiple times. One-shot network pruning is easy to reach the required sparsity, but the corresponding accuracy drop may be unacceptable with respect to different goals. On the other hand, iterative network pruning trims and retrains the network iteratively to maintain the accuracy, but suffering from the long runtime of this repetitive procedure. In this work, we propose an efficient approach to network pruning by removing redundant trainings. Experimental results show that our approach reduces 25\% to almost 60\% of training time with comparable network accuracy as compared to the state-of-the-art.
中文摘要--------------------------------------------------i
Abstract-------------------------------------------------ii
Acknowledgement-----------------------------------------iii
Contents-------------------------------------------------iv
List of Tables-------------------------------------------vi
List of Figures-----------------------------------------vii
1 Introduction--------------------------------------------1
2 Preliminaries-------------------------------------------4
2.1 Network Pruning---------------------------------------4
2.2 Magnitude-Based Weight Pruning------------------------5
2.3 Retraining Techniques---------------------------------5
2.4 One-Shot Pruning and Iterative Pruning----------------6
3 Proposed Approach---------------------------------------7
3.1 High Pruning Ratio for the First Iteration------------7
3.2 Take an Extra Step Back to a Resilient Point----------9
4 Experimental Results-----------------------------------11
4.1 ResNets on CIFAR-10----------------------------------12
4.2 VGG-16 on CIFAR-10-----------------------------------13
4.3 ResNets on CIFAR-100---------------------------------13
4.4 R(2+1)D-18 on EgoGesture-----------------------------14
5 Conclusion---------------------------------------------16
[1] Cao, C., Zhang, Y., Wu, Y., Lu, H. & Cheng, J. Egocentric gesture recognition
using recurrent 3d convolutional neural networks with spatiotemporal
transformer modules. Proceedings Of The IEEE International Conference On
Computer Vision. pp. 3763-3771 (2017)
[2] Frankle, J. & Carbin, M. The lottery ticket hypothesis: Finding sparse, trainable
neural networks. ArXiv Preprint ArXiv:1803.03635. (2018)
[3] Frankle, J., Dziugaite, G., Roy, D. & Carbin, M. Linear mode connectivity and
the lottery ticket hypothesis. International Conference On Machine Learning.
pp. 3259-3269 (2020)
[4] Fu, C. PyTorch-VGG-CIFAR10. (https://github.com/chengyangfu/pytorchvgg-
cifar10), Accessed: 2023-06-23
[5] Ghodasara, K. Overview of Decision Tree Pruning in Machine Learning. International
Research Journal Of Engineering And Technology (IRJET). 8 (2021)
[6] Han, S., Pool, J., Tran, J. & Dally, W. Learning both weights and connections
for efficient neural network. Advances In Neural Information Processing Systems.
28 (2015)
[7] He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition.
Proceedings Of The IEEE Conference On Computer Vision And Pattern
Recognition. pp. 770-778 (2016)
[8] He, Y., Kang, G., Dong, X., Fu, Y. & Yang, Y. Soft filter pruning for accelerating
deep convolutional neural networks. ArXiv Preprint ArXiv:1808.06866.
(2018)
[9] Idelbayev, Y. Proper ResNet Implementation for CIFAR10/CIFAR100 in
PyTorch. (https://github.com/akamaster/pytorch resnet cifar10), Accessed:
2023-05-14
[10] Krizhevsky, A., Hinton, G. & Others Learning multiple layers of features from
tiny images. (Toronto, ON, Canada,2009)
[11] Le, D. & Hua, B. Network pruning that matters: A case study on retraining
variants. ArXiv Preprint ArXiv:2105.03193. (2021)
[12] LeCun, Y., Denker, J. & Solla, S. Optimal brain damage. Advances In Neural
Information Processing Systems. 2 (1989)
[13] Li, H., Kadav, A., Durdanovic, I., Samet, H. & Graf, H. Pruning filters for
efficient convnets. ArXiv Preprint ArXiv:1608.08710. (2016)
[14] Liu, Z., Sun, M., Zhou, T., Huang, G. & Darrell, T. Rethinking the value of
network pruning. ArXiv Preprint ArXiv:1810.05270. (2018)
[15] Neyshabur, B., Li, Z., Bhojanapalli, S., LeCun, Y. & Srebro, N. Towards understanding
the role of over-parametrization in generalization of neural networks.
ArXiv Preprint ArXiv:1805.12076. (2018)
[16] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen,
T., Lin, Z., Gimelshein, N., Antiga, L. & Others Pytorch: An imperative style,
high-performance deep learning library. Advances In Neural Information Processing
Systems. 32 (2019)
[17] Renda, A., Frankle, J. & Carbin, M. Comparing rewinding and fine-tuning in
neural network pruning. ArXiv Preprint ArXiv:2003.02389. (2020)
[18] Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale
image recognition. ArXiv Preprint ArXiv:1409.1556. (2014)
[19] Smith, L. & Topin, N. Super-convergence: Very fast training of neural networks
using large learning rates. Artificial Intelligence And Machine Learning For
Multi-domain Operations Applications. 11006 pp. 369-386 (2019)
[20] Tran, D., Wang, H., Torresani, L., Ray, J., LeCun, Y. & Paluri, M. A closer
look at spatiotemporal convolutions for action recognition. Proceedings Of The
IEEE Conference On Computer Vision And Pattern Recognition. pp. 6450-6459
(2018)
[21] Zhang, Y., Cao, C., Cheng, J. & Lu, H. EgoGesture: A new dataset and benchmark
for egocentric hand gesture recognition. IEEE Transactions On Multimedia.
20, 1038-1050 (2018)
 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
* *