Skip to main navigation Skip to search Skip to main content

Dictionary Pair-based Data-Free Fast Deep Neural Network Compression

Yangcheng Gao, Zhao Zhang*, Haijun Zhang, Mingbo Zhao, Yi Yang, Meng Wang

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

Deep neural network (DNN) compression can reduce the memory footprint of deep networks effectively, so that the deep model can be deployed on the portable devices. However, most of the existing model compression methods cost lots of time, e.g., vector quantization or pruning, which makes them inept to the real-world applications that need fast online computation. In this paper, we therefore explore how to accelerate the model compression process by reducing the computation cost. Then, we propose a new deep model compression method, termed Dictionary Pair-based Data-Free Fast DNN Compression, which aims at reducing the memory consumption of DNNs without extra training and can greatly improve the compression efficiency. Specifically, our proposed method performs tensor decomposition on the DNN model with a fast dictionary pair learning-based reconstruction approach, which can be deployed on different layers (e.g., convolution and fully-connection layers). Given a pre-trained DNN model, we first divide the parameters (i.e., weights) of each layer into a series of partitions for dictionary pair-based fast reconstruction, which can potentially discover more fine-grained information and provide the possibility for parallel model compression. Then, dictionaries of less memory occupation are learned to reconstruct the weights. Extensive experiments on popular DNNs (i.e., VGG-16, ResNet-18 and ResNet-50) showed that our proposed weight compression method can significantly reduce the memory footprint and speed up the compression process, with less performance loss.
Original languageEnglish
Title of host publicationProceedings - 21st IEEE International Conference on Data Mining, ICDM 2021
EditorsJames Bailey, Pauli Miettinen, Yun Sing Koh, Dacheng Tao, Xindong Wu
PublisherIEEE
Pages121-130
ISBN (Electronic)9781665423984
ISBN (Print)9781665423991
DOIs
Publication statusPublished - Dec 2021
Event21st IEEE International Conference on Data Mining (ICDM 2021) - Virtual, Auckland, New Zealand
Duration: 7 Dec 202110 Dec 2021

Publication series

NameProceedings - IEEE International Conference on Data Mining, ICDM
Volume2021-December
ISSN (Print)1550-4786
ISSN (Electronic)2374-8486

Conference

Conference21st IEEE International Conference on Data Mining (ICDM 2021)
Abbreviated titleIEEE ICDM 2021
PlaceNew Zealand
CityAuckland
Period7/12/2110/12/21

Research Keywords

  • dictionary pair-based fast compression of DNNs
  • fast weight reconstruction
  • less performance loss
  • Model compression efficiency

Fingerprint

Dive into the research topics of 'Dictionary Pair-based Data-Free Fast Deep Neural Network Compression'. Together they form a unique fingerprint.

Cite this