|
1. Dario Amodei, et al., “Deep speech 2: End-to-End Speech Recognition in English and Mandarin,” in International conference on machine learning, PMLR, 2016, pp. 173-182. 2. Sajid Anwar, et al., “Structured Pruning of Deep Convolutional Neural Networks,” in ACM Journal on Emerging Technologies in Computing Systems (JETC), 2017, vol. 13, no. 3, pp. 1-18. 3. Chia-Chih Chi and Jie-Hong R Jiang, “Logic Synthesis of Binarized Neural Networks for Efficient Circuit Implementation,” in Proc. of ICCAD, 2018. 4. Matthieu Courbariaux, et al., “Binaryconnect: Training Deep Neural Networks with Binary Weights during Propagations,” arXiv preprint arXiv:1511.00363, 2015. 5. Matthieu Courbariaux, et al., “Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1,” arXiv preprint arXiv:1602.02830, 2016. 6. Ya-Chun Chang, et al., “A Convolutional Result Sharing Approach for Binarized Neural Network Inference,” in Proc. of DATE, 2020, pp. 780-785. 7. Itay Hubara, et al., “Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations,” JMLR, 2015, vol. 18, no. 1, pp. 6869-6898. 8. Song Han, et al., “Deep compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman coding,” arXiv preprint arXiv:1510.00149, 2015. 9. Forrest N Iandola, et al., “SqueezeNet: AlexNet-level Accuracy with 50x Fewer Parameters and < 0.5 MB Model Size,” arXiv preprint arXiv:1602.07360, 2016. 10. Sergey Ioffe and Christian Szegedy, “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” in Proc. of ICML, 2018, pp. 448-456. 11. Alex Krizhevsky and Geoffrey Hinton, “Learning Multiple Layers of Features from Tiny Images,” Citeseer, 2009. 12. Alex Krizhevsky, et al., “Imagenet Classification with Deep Convolutional Neural Networks,” in Advances in neural information processing systems, 2012, vol. 25, pp. 1097-1105. 13. Yann LeCun, ”The MNIST Database of Handwritten Digits,” http://yann.lecun.com/exdb/mnist/, 1998. 14. Chigozie Nwankpa, et al., “Activation Functions: Comparison of Trends in Practice and Research for Deep Learning, ” arXiv preprint arXiv:1811.03378, 2018. 15. Sridhar Narayan, “The Generalized Sigmoid Activation Function: Competitive Supervised Learning,” Information Sciences, 1997, vol. 99, no. 1-2, pp. 69-82. 16. Jiantao Qiu, et al., “Going Deeper with Embedded FPGA Platform for Convolutional Neural Network,” in Proc. of the FPGA, 2016, pp. 26-35. 17. Mohammad Rastegari, et al., “Xnor-net: Imagenet Classification Using Binary Convolutional Neural Networks,” in Proc. of ECCV, Springer, 2016, pp. 525- 542. 18. Yaman Umuroglu, et al., “Finn: A Framework for Fast, Scalable Binarized Neural Network Inference,” in Proc. of FPGA, 2017, pp. 65-74. 19. Bing Xu, et al., “Empirical Evaluation of Rectified Activations in Convolutional Network,” arXiv preprint arXiv:1505.00853, 2015. 20. Tien-Ju Yang, et al., “Designing Energy-efficient Convolutional Neural Networks Using Energy-aware Pruning,” in Proc. of CVPR, 2017, pp. 5687-5695. 21. Shuchang Zhou, et al., “Dorefa-net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradient,” arXiv preprint arXiv:1606.06160, 2016. 22. Ye Zhang and Byron Wallace, “A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification,” arXiv preprint arXiv:1510.03820, 2015.
|