IEICE Transactions on Electronics
Online ISSN : 1745-1353
Print ISSN : 0916-8524

この記事には本公開記事があります。本公開記事を参照してください。
引用する場合も本公開記事を引用してください。

Weight compression MAC accelerator for effective inference of deep learning
Asuka MakiDaisuke MiyashitaShinichi SasakiKengo NakataFumihiko TachibanaTomoya SuzukiJun DeguchiRyuichi Fujimoto
著者情報
ジャーナル 認証あり 早期公開

論文ID: 2019CTP0007

この記事には本公開記事があります。
詳細
抄録

Many studies of deep neural networks have reported inference accelerators for improved energy efficiency. We propose methods for further improving energy efficiency while maintaining recognition accuracy, which were developed by the co-design of a filter-by-filter quantization scheme with variable bit precision and a hardware architecture that fully supports it. Filter-wise quantization reduces the average bit precision of weights, so execution times and energy consumption for inference are reduced in proportion to the total number of computations multiplied by the average bit precision of weights. The hardware utilization is also improved by a bit-parallel architecture suitable for granularly quantized bit precision of weights. We implement the proposed architecture on an FPGA and demonstrate that the execution cycles are reduced to 1/5.3 for ResNet-50 on ImageNet in comparison with a conventional method, while maintaining recognition accuracy.

著者関連情報
© 2020 The Institute of Electronics, Information and Communication Engineers
feedback
Top