Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Semantic Regularisation for Recurrent Image Annotation

About

The "CNN-RNN" design pattern is increasingly widely applied in a variety of image annotation tasks including multi-label classification and captioning. Existing models use the weakly semantic CNN hidden layer or its transform as the image embedding that provides the interface between the CNN and RNN. This leaves the RNN overstretched with two jobs: predicting the visual concepts and modelling their correlations for generating structured annotation output. Importantly this makes the end-to-end training of the CNN and RNN slow and ineffective due to the difficulty of back propagating gradients through the RNN to train the CNN. We propose a simple modification to the design pattern that makes learning more effective and efficient. Specifically, we propose to use a semantically regularised embedding layer as the interface between the CNN and RNN. Regularising the interface can partially or completely decouple the learning problems, allowing each to be more effectively trained and jointly training much more efficient. Extensive experiments show that state-of-the art performance is achieved on multi-label classification as well as image captioning.

Feng Liu, Tao Xiang, Timothy M. Hospedales, Wankou Yang, Changyin Sun• 2016

Related benchmarks

TaskDatasetResultRank
Multi-Label ClassificationNUS-WIDE (test)--
112
Multi-label Image ClassificationNUS-WIDE 81 concept labels (test)
F1 (Class)52.8
29
Multi-label Image ClassificationMS-COCO (test)--
24
Multi-label Image ClassificationMS-COCO 2014 (test)--
24
Showing 4 of 4 rows

Other info

Follow for update