IEICE Electronics Express
Online ISSN : 1349-2543
ISSN-L : 1349-2543
LETTER
HFOD: A hardware-friendly quantization method for object detection on embedded FPGAs
Fei ZhangZiyang GaoJiaming HuangPeining ZhenHai-Bao ChenJie Yan
Author information
JOURNAL FREE ACCESS

2022 Volume 19 Issue 8 Pages 20220067

Details
Abstract

There are two research hotspots for improving performance and energy efficiency of the inference phase of Convolutional neural networks (CNNs). The first one is model compression techniques while the second is hardware accelerator implementation. To overcome the incompatibility of algorithm optimization and hardware design, this paper proposes HFOD, a hardware-friendly quantization method for object detection on embedded FPGAs. We adopt a channel-wise, uniform quantization method to compress YOLOv3-Tiny model. Weights are quantized to 2-bit while activations are quantized to 8-bit for all convolutional layers. To achieve highly-efficient implementations on FPGA, we add batch normalization (BN) layer fusion in quantization process. A flexible, efficient convolutional unit structure is designed to utilize hardware-friendly quantization, and the accelerator is developed based on an automatic synthesis template. Experimental results show that the resources of FPGA in the proposed accelerator design contribute more computing performance compared with regular 8-bit/16-bit fixed point quantization. The model size and the activation size of the proposed network with 2-bit weights and 8-bit activations can be effectively reduced by 16× and 4× with a small amount of accuracy loss, respectively. Our HFOD method can achieve 90.6 GOPS on PYNQ-Z2 at 150MHz, which is 1.4× faster and 2× better in power efficiency than peer FPGA implementation on the same platform.

Content from these authors
© 2022 by The Institute of Electronics, Information and Communication Engineers
Previous article Next article
feedback
Top