Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Zero-Cost Proxies for Lightweight NAS

About

Neural Architecture Search (NAS) is quickly becoming the standard methodology to design neural network models. However, NAS is typically compute-intensive because multiple models need to be evaluated before choosing the best one. To reduce the computational power and time needed, a proxy task is often used for evaluating each model instead of full training. In this paper, we evaluate conventional reduced-training proxies and quantify how well they preserve ranking between multiple models during search when compared with the rankings produced by final trained accuracy. We propose a series of zero-cost proxies, based on recent pruning literature, that use just a single minibatch of training data to compute a model's score. Our zero-cost proxies use 3 orders of magnitude less computation but can match and even outperform conventional proxies. For example, Spearman's rank correlation coefficient between final validation accuracy and our best zero-cost proxy on NAS-Bench-201 is 0.82, compared to 0.61 for EcoNAS (a recently proposed reduced-training proxy). Finally, we use these zero-cost proxies to enhance existing NAS search algorithms such as random search, reinforcement learning, evolutionary search and predictor-based search. For all search methodologies and across three different NAS datasets, we are able to significantly improve sample efficiency, and thereby decrease computation, by using our zero-cost proxies. For example on NAS-Bench-101, we achieved the same accuracy 4$\times$ quicker than the best previous result. Our code is made public at: https://github.com/mohsaied/zero-cost-nas.

Mohamed S. Abdelfattah, Abhinav Mehrotra, {\L}ukasz Dudziak, Nicholas D. Lane• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy75.9
3518
Image ClassificationCIFAR-10 (test)
Accuracy95.1
3381
Image ClassificationCIFAR-10 (test)
Accuracy (Clean)92.95
273
Neural Architecture SearchNAS-Bench-201 ImageNet-16-120 (test)
Accuracy43.64
86
Neural Architecture SearchCIFAR-10 NAS-Bench-201 (val)
Accuracy90.19
86
Neural Architecture SearchNASBench-201 CIFAR-10
Retrieving Rate @ Top 10%52
85
Neural Architecture SearchNAS-Bench-201 CIFAR-10 (test)
Accuracy93.45
85
Neural Architecture SearchImageNet16-120 NAS-Bench-201 (val)
Accuracy43.24
79
Neural Architecture SearchNAS-Bench-201 CIFAR-100 (test)
Accuracy70.73
78
Neural Architecture SearchCIFAR-100 NAS-Bench-201 (val)
Accuracy70.55
67
Showing 10 of 63 rows

Other info

Follow for update