Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer

About

Attention plays a critical role in human visual experience. Furthermore, it has recently been demonstrated that attention can also play an important role in the context of applying artificial neural networks to a variety of tasks from fields such as computer vision and NLP. In this work we show that, by properly defining attention for convolutional neural networks, we can actually use this type of information in order to significantly improve the performance of a student CNN network by forcing it to mimic the attention maps of a powerful teacher network. To that end, we propose several novel methods of transferring attention, showing consistent improvement across a variety of datasets and convolutional neural network architectures. Code and models for our experiments are available at https://github.com/szagoruyko/attention-transfer

Sergey Zagoruyko, Nikos Komodakis• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy74.08
3518
Image ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy70.72
1866
Image ClassificationImageNet-1k (val)
Top-1 Accuracy71.03
1453
Image ClassificationImageNet (val)
Top-1 Acc71.03
1206
Image ClassificationCIFAR-10 (test)
Accuracy86.86
906
Image ClassificationCIFAR-100 (val)--
661
Image ClassificationCIFAR-100
Top-1 Accuracy78.64
622
Image ClassificationCIFAR100 (test)
Top-1 Accuracy75.28
377
Image ClassificationTinyImageNet (test)
Accuracy34.44
366
Image ClassificationSTL-10 (test)
Accuracy67.37
357
Showing 10 of 64 rows

Other info

Code

Follow for update