Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

A Survey of Model Compression and Acceleration for Deep Neural Networks

About

Deep neural networks (DNNs) have recently achieved great success in many visual recognition tasks. However, existing deep neural network models are computationally expensive and memory intensive, hindering their deployment in devices with low memory resources or in applications with strict latency requirements. Therefore, a natural thought is to perform model compression and acceleration in deep networks without significantly decreasing the model performance. During the past five years, tremendous progress has been made in this area. In this paper, we review the recent techniques for compacting and accelerating DNN models. In general, these techniques are divided into four categories: parameter pruning and quantization, low-rank factorization, transferred/compact convolutional filters, and knowledge distillation. Methods of parameter pruning and quantization are described first, after that the other techniques are introduced. For each category, we also provide insightful analysis about the performance, related applications, advantages, and drawbacks. Then we go through some very recent successful methods, for example, dynamic capacity networks and stochastic depths networks. After that, we survey the evaluation matrices, the main datasets used for evaluating the model performance, and recent benchmark efforts. Finally, we conclude this paper, discuss remaining the challenges and possible directions for future work.

Yu Cheng, Duo Wang, Pan Zhou, Tao Zhang• 2017

Related benchmarks

TaskDatasetResultRank
Uncertainty QuantificationCIFAR10
NLL0.277
42
Uncertainty QuantificationFashionMNIST
NLL0.283
42
Uncertainty QuantificationImageNet-10
NLL0.276
42
Image ClassificationCIFAR-10-C (test)
NLL0.341
28
Image ClassificationCIFAR-10-C gaussian blur
NLL0.352
6
Image ClassificationCIFAR-10-C brightness
NLL0.34
6
Image ClassificationCIFAR-10-C elastic transform
NLL0.522
6
Image ClassificationCIFAR-10-C impulse noise
NLL1.836
6
Uncertainty QuantificationMNIST-C (test)
NLL (brightness)0.102
6
Showing 9 of 9 rows

Other info

Follow for update