InterActive: Inter-Layer Activeness Propagation
About
An increasing number of computer vision tasks can be tackled with deep features, which are the intermediate outputs of a pre-trained Convolutional Neural Network. Despite the astonishing performance, deep features extracted from low-level neurons are still below satisfaction, arguably because they cannot access the spatial context contained in the higher layers. In this paper, we present InterActive, a novel algorithm which computes the activeness of neurons and network connections. Activeness is propagated through a neural network in a top-down manner, carrying high-level context and improving the descriptive power of low-level and mid-level neurons. Visualization indicates that neuron activeness can be interpreted as spatial-weighted neuron responses. We achieve state-of-the-art classification performance on a wide range of image datasets.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Fine-grained Image Classification | Oxford Flowers | Accuracy96.4 | 49 | |
| Fine-grained Image Classification | Oxford-IIIT Pets | Accuracy93.5 | 29 | |
| Fine-grained Image Classification | Oxford 102 Flowers (test) | Accuracy96.4 | 19 | |
| Fine-grained Visual Categorization | Oxford-IIIT Pets (test) | Accuracy93.5 | 6 |