Selective Pruning of Sparsity-Supported Energy-Efficient Accelerator for Convolutional Neural Networks

Chia Chi Liu*, I. Chyn Wey, Xuezhi Zhang, T. Hui Teo

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Convolutional Neural Networks (CNNs) are widely used in various fields with the rapid development of Deep Learning (DL). However, the massive amount of parameters and huge models severely limit the calculation speed and performance of the model. To address this, model compression has become one of the most popular methods, and quantization and pruning are two of the most common techniques. Quantization reduces the bit width in exchange for lower power consumption and computing time, while pruning reduces the number of parameters to reduce memory access and operation time. However, the existing hardware architecture is still unable to satisfy real-time and low power consumption requirements simultaneously. Therefore, scholars have been paying attention to Application Specific Integrated Circuits (ASICs) and low-bit quantization co-optimization. This paper aims to co-design sparsity and quantization energy-saving accelerators. It uses the characteristics of selective pruning to design low-cost sparsity hardware and employs low-cost comparison circuits to solve the problem of energy-saving data flow dependencies and improve traditional low-efficiency non-real-time methods. In terms of hardware, compared to Dynamic Region-based Quantization (DRQ), one of the state-of-the-art hardware accelerators, the proposed network reduces the area by 43.22%, power consumption by 52.17%, and computing time by 25.36%. Additionally, it increases power efficiency by 1.29 times.

Original languageEnglish
Title of host publicationProceedings - 2023 16th IEEE International Symposium on Embedded Multicore/Many-Core Systems-on-Chip, MCSoC 2023
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages454-461
Number of pages8
ISBN (Electronic)9798350393613
DOIs
StatePublished - 2023
Event16th IEEE International Symposium on Embedded Multicore/Many-Core Systems-on-Chip, MCSoC 2023 - Singapore, Singapore
Duration: 18 12 202321 12 2023

Publication series

NameProceedings - 2023 16th IEEE International Symposium on Embedded Multicore/Many-Core Systems-on-Chip, MCSoC 2023

Conference

Conference16th IEEE International Symposium on Embedded Multicore/Many-Core Systems-on-Chip, MCSoC 2023
Country/TerritorySingapore
CitySingapore
Period18/12/2321/12/23

Bibliographical note

Publisher Copyright:
© 2023 IEEE.

Keywords

  • Convolutional Neural Network
  • Hardware Accelerator
  • Low Power Consumption
  • Model Compression
  • Pruning
  • Quantization
  • Sparsity

Fingerprint

Dive into the research topics of 'Selective Pruning of Sparsity-Supported Energy-Efficient Accelerator for Convolutional Neural Networks'. Together they form a unique fingerprint.

Cite this