Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CAP: A Context-Aware Neural Predictor for NAS

About

Neural predictors are effective in boosting the time-consuming performance evaluation stage in neural architecture search (NAS), owing to their direct estimation of unseen architectures. Despite the effectiveness, training a powerful neural predictor with fewer annotated architectures remains a huge challenge. In this paper, we propose a context-aware neural predictor (CAP) which only needs a few annotated architectures for training based on the contextual information from the architectures. Specifically, the input architectures are encoded into graphs and the predictor infers the contextual structure around the nodes inside each graph. Then, enhanced by the proposed context-aware self-supervised task, the pre-trained predictor can obtain expressive and generalizable representations of architectures. Therefore, only a few annotated architectures are sufficient for training. Experimental results in different search spaces demonstrate the superior performance of CAP compared with state-of-the-art neural predictors. In particular, CAP can rank architectures precisely at the budget of only 172 annotated architectures in NAS-Bench-101. Moreover, CAP can help find promising architectures in both NAS-Bench-101 and DARTS search spaces on the CIFAR-10 dataset, serving as a useful navigator for NAS to explore the search space efficiently.

Han Ji, Yuqi Feng, Yanan Sun• 2024

Related benchmarks

TaskDatasetResultRank
Neural Architecture SearchCIFAR-10 NAS-Bench-201 (val)
Accuracy94.34
86
Neural Architecture SearchNASBench-201 ImageNet16-120
Rank3
38
Neural Architecture SearchNAS-Bench-101
Accuracy94.18
19
Neural Architecture SearchNASBench-201 cifar100
Rank4
19
Neural Architecture SearchNAS-Bench-201 CIFAR-100
Accuracy73.41
19
Neural Architecture SearchNASBench-201 cifar10 (val)
Rank6
19
Neural Architecture SearchNASBench-101
Rank8
19
Showing 7 of 7 rows

Other info

Follow for update