研究生: |
蔡宥璿 Tsai, You-Hsuen |
---|---|
論文名稱: |
藉由找出過濾器間相同的部分並分享運算結果來最小化二值化神經網路中所需運算的研究 Minimizing Computation in Binarized Neural Network Inference using Partial-Filter Sharing |
指導教授: |
王俊堯
Wang, Chun-Yao |
口試委員: |
陳勇志
Chen, Yung-Chih 陳聿廣 Chen, Yu-Guan |
學位類別: |
碩士 Master |
系所名稱: |
電機資訊學院 - 資訊工程學系 Computer Science |
論文出版年: | 2022 |
畢業學年度: | 110 |
語文別: | 英文 |
論文頁數: | 24 |
中文關鍵詞: | 二值化神經網路 、過濾器重複特性 |
外文關鍵詞: | Binarized Neural Networks, Filter Repetitions |
相關次數: | 點閱:1 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
二值化神經網路是卷積式神經網路的一種變形,其不同之處在於二值化神經網路
中的神經元是二進制的權重和二進制的輸出。此種網路近年來成為了在資源有限
的設備上部署人工智慧的一種具前景的方法。由於二值化權重的特性,二值化神
經網路在計算上不再需要乘法器,並且過濾器之間也有著較高程度的相似性。在
這篇論文中,我們提出了一種利用部分過濾器間重複的特性的方法,並將此方法
與目前最先進的技術做結合,以減少在現場可程式化邏輯閘陣列(FPGA)上實現
所需的硬體成本和合成時間。
Binarized Neural Network (BNN), which is a variant of Convolutional Neural Network (CNN) with binary weights and binary outputs on a neuron, has emerged as a promising approach to deploy artificial intelligence on resource-restricted devices in recent years. Due to the binarized weights, there are no longer required multipliers for computation, and there exist relatively high similarities among filters as well. In this work, we propose a partial-filter sharing approach and integrate it with the state-of-the-art to reduce the hardware cost and the synthesis time onto Field Programmable Gate Arrays (FPGAs). Experimental results show the effectiveness of our approach in terms of look-up table (LUT) counts and synthesis time on an FPGA platform. As compared to the state-of-the-art, the LUT reduction rate by our approach is 42.6% on average without any accuracy loss, and 38.9% synthesis time can be also saved on average.
[1] Chia-Chih Chi and Jie-Hong R Jiang, “Logic Synthesis of Binarized Neural Networks for Efficient Circuit Implementation,” in Proc. of ICCAD, 2018.
[2] Matthieu Courbariaux, et al., “Binaryconnect: Training Deep Neural Networks with Binary Weights during Propagations,” arXiv preprint arXiv:1511.00363, 2015.
[3] Matthieu Courbariaux, et al., “Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1,” arXiv preprint arXiv:1602.02830, 2016.
[4] Ya-Chun Chang, et al., “A Convolutional Result Sharing Approach for Binarized Neural Network Inference,” in Proc. of DATE, 2020, pp. 780-785.
[5] Tong Geng, et al., “O3BNN: An Out-Of-Order Architecture for High-
Performance Binarized Neural Network Inference with Fine-Grained Pruning,”in Proc. of ICS, 2019, pp. 461-472.
[6] Itay Hubara, et al., “Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations,” JMLR, 2015, vol. 18, no. 1, pp.6869-6898.
[7] Alex Krizhevsky and Geoffrey Hinton, “Learning Multiple Layers of Features from Tiny Images,” Citeseer, 2009.
[8] Qingliang Liu, et al., “TCP-Net: Minimizing Operation Counts of Binarized Neural Network Inference,” in Proc. of ISCAS, 2021, pp. 1-5.
9] Yann LeCun, et al., “Gradient-based learning applied to document recognition,”
in Proc. of IEEE , Nov. 1998, pp. 2278-2324.
[10] Ya Le and Xuan S. Yang, “Tiny ImageNet Visual Recognition Challenge,” 2015.
[11] Mohammad Rastegari, et al., “Xnor-net: Imagenet Classification Using Binary Convolutional Neural Networks,” in Proc. of ECCV, Springer, 2016, pp. 525-542.
[12] Richard L Rudell, “Logic Synthesis for VLSI Design,” PHD dissertation, University of California, Berkeley, 1989.
[13] Karen Simonyan and Andrew Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition“ in International Conference on Learning Representations, 2015.
[14] Wei Tang, et al., “How to train a compact binary neural network with high accuracy?,” in Proc. of AAI, 2017, pp. 2625-2631.
[15] Yaman Umuroglu, et al., “Finn: A Framework for Fast, Scalable Binarized Neural Network Inference,” in Proc. of FPGA, 2017, pp. 65-74.
[16] Sheng-Hsiu Wei, et al., “A Flexible Result Sharing Approach to Binarized Neural Networks Optimization,” MS thesis, University of Tsing Hua, 2021.
[17] Shuchang Zhou, et al., “Dorefa-net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradient,” arXiv preprint arXiv:1606.06160, 2016.