Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Distilling Knowledge via Knowledge Review

About

Knowledge distillation transfers knowledge from the teacher network to the student one, with the goal of greatly improving the performance of the student network. Previous methods mostly focus on proposing feature transformation and loss functions between the same level's features to improve the effectiveness. We differently study the factor of connection path cross levels between teacher and student networks, and reveal its great importance. For the first time in knowledge distillation, cross-stage connection paths are proposed. Our new review mechanism is effective and structurally simple. Our finally designed nested and compact framework requires negligible computation overhead, and outperforms other methods on a variety of tasks. We apply our method to classification, object detection, and instance segmentation tasks. All of them witness significant student network performance improvement. Code is available at https://github.com/Jia-Research-Lab/ReviewKD

Pengguang Chen, Shu Liu, Hengshuang Zhao, Jiaya Jia• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy77.91
3518
Object DetectionCOCO 2017 (val)
AP40.36
2454
Image ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy72.56
1866
Image ClassificationImageNet-1k (val)
Top-1 Accuracy72.56
1453
Image ClassificationImageNet (val)
Top-1 Acc72.56
1206
Instance SegmentationCOCO 2017 (val)
APm0.3957
1144
Image ClassificationImageNet-1k (val)
Top-1 Accuracy72.56
840
Image ClassificationCIFAR-100 (val)
Accuracy77.78
661
Image ClassificationCIFAR-100
Top-1 Accuracy75.63
622
Image ClassificationCIFAR-10--
507
Showing 10 of 25 rows

Other info

Code

Follow for update