Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Zero Time Waste: Recycling Predictions in Early Exit Neural Networks

About

The problem of reducing processing time of large deep learning models is a fundamental challenge in many real-world applications. Early exit methods strive towards this goal by attaching additional Internal Classifiers (ICs) to intermediate layers of a neural network. ICs can quickly return predictions for easy examples and, as a result, reduce the average inference time of the whole model. However, if a particular IC does not decide to return an answer early, its predictions are discarded, with its computations effectively being wasted. To solve this issue, we introduce Zero Time Waste (ZTW), a novel approach in which each IC reuses predictions returned by its predecessors by (1) adding direct connections between ICs and (2) combining previous outputs in an ensemble-like manner. We conduct extensive experiments across various datasets and architectures to demonstrate that ZTW achieves a significantly better accuracy vs. inference time trade-off than other recently proposed early exit methods.

Maciej Wo{\l}czyk, Bartosz W\'ojcik, Klaudia Ba{\l}azy, Igor Podolak, Jacek Tabor, Marek \'Smieja, Tomasz Trzci\'nski• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy76.4
3518
Image ClassificationCIFAR-10 (test)
Accuracy94.7
3381
Image ClassificationTinyImageNet (test)
Accuracy60.5
366
Image ClassificationImageNet (test)--
235
Retinal Disease ClassificationOCT 2017 (test)
Accuracy98.5
24
Showing 5 of 5 rows

Other info

Code

Follow for update