An Efficient Implementation of Convolutional Neural Network With CLIP-Q Quantization on FPGA

Wei Cheng, Ing Chao Lin, Yun Yang Shih

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Convolutional neural networks (CNNs) have achieved tremendous success in the computer vision domain recently. The pursue for better model accuracy drives the model size and the storage requirements of CNNs as well as the computational complexity. Therefore, Compression Learning by InParallel Pruning-Quantization (CLIP-Q) was proposed to reduce a vast amount of weight storage requirements by using a few quantized segments to represent all weights in a CNN layer. Among various quantization strategies, CLIP-Q is suitable for hardware accelerators because it reduces model size significantly while maintaining the full-precision model accuracy. However, the current CLIP-Q approach did not consider the hardware characteristics and it is not straightforward when mapped to a CNN hardware accelerator. In this work, we propose a software-hardware codesign platform that includes a modified version of CLIP-Q algorithm and a hardware accelerator, which consists of 5×5 reconfigurable convolutional arrays with input and output channel parallelization. Additionally, the proposed CNN accelerator maintains the same accuracy of a full-precision CNN in Cifar-10 and Cifar-100 datasets.

Original languageEnglish
Pages (from-to)4093-4102
Number of pages10
JournalIEEE Transactions on Circuits and Systems I: Regular Papers
Volume69
Issue number10
DOIs
Publication statusPublished - 2022 Oct 1

All Science Journal Classification (ASJC) codes

  • Hardware and Architecture
  • Electrical and Electronic Engineering

Fingerprint

Dive into the research topics of 'An Efficient Implementation of Convolutional Neural Network With CLIP-Q Quantization on FPGA'. Together they form a unique fingerprint.

Cite this