Hardware-Friendly Progressive Pruning Framework for CNN Model Compression using Universal Pattern Sets

Wei Cheng Chou, Cheng Wei Huang, Juinn Dar Huang

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Scopus citations

Abstract

Pattern-based weight pruning on CNNs has been proven an effective model reduction technique. In this paper, we first present how to select hardware-friendly pruning pattern sets that are universal to various models. We then propose a progressive pruning framework, which produces more globally optimized outcomes. Moreover, to the best of our knowledge, this is the first paper dealing with the pruning issue of the first and also the most sensitive layer of a CNN model through a two-staged pruning strategy. Experiment results show that the proposed framework achieves 2.25x/2x computation/model reduction while minimizing the accuracy loss.

Original languageEnglish
Title of host publication2022 International Symposium on VLSI Design, Automation and Test, VLSI-DAT 2022 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781665409216
DOIs
StatePublished - 2022
Event2022 International Symposium on VLSI Design, Automation and Test, VLSI-DAT 2022 - Hsinchu, Taiwan
Duration: 18 Apr 202221 Apr 2022

Publication series

Name2022 International Symposium on VLSI Design, Automation and Test, VLSI-DAT 2022 - Proceedings

Conference

Conference2022 International Symposium on VLSI Design, Automation and Test, VLSI-DAT 2022
Country/TerritoryTaiwan
CityHsinchu
Period18/04/2221/04/22

Keywords

  • convolutional neural network (CNN)
  • first layer pruning
  • model compression
  • pattern pruning
  • universal pattern sets
  • weight pruning

Fingerprint

Dive into the research topics of 'Hardware-Friendly Progressive Pruning Framework for CNN Model Compression using Universal Pattern Sets'. Together they form a unique fingerprint.

Cite this