Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AMC: AutoML for Model Compression and Acceleration on Mobile Devices

About

Model compression is a critical technique to efficiently deploy neural network models on mobile devices which have limited computation resources and tight power budgets. Conventional model compression techniques rely on hand-crafted heuristics and rule-based policies that require domain experts to explore the large design space trading off among model size, speed, and accuracy, which is usually sub-optimal and time-consuming. In this paper, we propose AutoML for Model Compression (AMC) which leverage reinforcement learning to provide the model compression policy. This learning-based compression policy outperforms conventional rule-based compression policy by having higher compression ratio, better preserving the accuracy and freeing human labor. Under 4x FLOPs reduction, we achieved 2.7% better accuracy than the handcrafted model compression policy for VGG-16 on ImageNet. We applied this automated, push-the-button compression pipeline to MobileNet and achieved 1.81x speedup of measured inference latency on an Android phone and 1.43x speedup on the Titan XP GPU, with only 0.1% loss of ImageNet Top-1 accuracy.

Yihui He, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, Song Han• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 (test)--
3381
Image ClassificationCIFAR-10 (test)
Accuracy91.9
906
Image ClassificationCIFAR-10--
471
Image ClassificationImageNet
Top-1 Accuracy70.8
429
Image ClassificationImageNet (val)
Top-1 Accuracy70.5
76
Image ClassificationImageNet (val)
Top-1 Accuracy70.5
68
Image ClassificationImageNet
Top-1 Accuracy71.8
60
Image ClassificationImageNet (val)
T170.8
45
Image ClassificationCIFAR-10
Accuracy (Pruned)91.9
33
Image ClassificationImageNet-1k (val)
Accuracy70.49
23
Showing 10 of 16 rows

Other info

Follow for update