Skip to main navigation Skip to search Skip to main content

Novel CNN-based AP2D-net accelerator: An area and power efficient solution for real-time applications on mobile FPGA

  • Shuai Li
  • , Kuangyuan Sun
  • , Yukui Luo
  • , Nandakishor Yadav
  • , Ken Choi

Research output: Contribution to journalArticlepeer-review

10 Scopus citations

Abstract

Standard convolutional neural networks (CNNs) have large amounts of data redundancy, and the same accuracy can be obtained even in lower bit weights instead of floating-point representation. Most CNNs have to be developed and executed on high-end GPU-based workstations, for which it is hard to transplant the existing implementations onto portable edge FPGAs because of the limitation of on-chip block memory storage size and battery capacity. In this paper, we present adaptive pointwise convolution and 2D convolution joint network (AP2D-Net), an ultra-low power and relatively high throughput system combined with dynamic precision weights and activation. Our system has high performance, and we make a trade-off between accuracy and power efficiency by adopting unmanned aerial vehicle (UAV) object detection scenarios. We evaluate our system on the Zynq UltraScale+ MPSoC Ultra96 mobile FPGA platform. The target board can get the real-time speed of 30 fps under 5.6 W, and the FPGA on-chip power is only 0.6 W. The power efficiency of our system is 2.8× better than the best system design on a Jetson TX2 GPU and 1.9× better than the design on a PYNQ-Z1 SoC FPGA.

Original languageEnglish
Article number832
JournalElectronics (Switzerland)
Volume9
Issue number5
DOIs
StatePublished - May 2020

Keywords

  • Binary neural network
  • Deep neural network accelerator
  • FPGA
  • Object detection
  • Parallel computing
  • Pipeline architecture
  • Power efficiency
  • UAV

Fingerprint

Dive into the research topics of 'Novel CNN-based AP2D-net accelerator: An area and power efficient solution for real-time applications on mobile FPGA'. Together they form a unique fingerprint.

Cite this