Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Robust Visual-Semantic Embeddings

About

Many of the existing methods for learning joint embedding of images and text use only supervised information from paired images and its textual attributes. Taking advantage of the recent success of unsupervised learning in deep neural networks, we propose an end-to-end learning framework that is able to extract more robust multi-modal representations across domains. The proposed method combines representation learning models (i.e., auto-encoders) together with cross-domain learning criteria (i.e., Maximum Mean Discrepancy loss) to learn joint embeddings for semantic and visual features. A novel technique of unsupervised-data adaptation inference is introduced to construct more comprehensive embeddings for both labeled and unlabeled data. We evaluate our method on Animals with Attributes and Caltech-UCSD Birds 200-2011 dataset with a wide range of applications, including zero and few-shot image recognition and retrieval, from inductive to transductive settings. Empirically, we show that our framework improves over the current state of the art on many of the considered tasks.

Yao-Hung Hubert Tsai, Liang-Kang Huang, Ruslan Salakhutdinov• 2017

Related benchmarks

TaskDatasetResultRank
Action RecognitionNTU RGB+D 60 (X-sub)
Accuracy75.37
467
Few-shot classificationtieredImageNet (test)--
282
Generalized Zero-Shot LearningCUB
H Score32.3
250
Few-shot Image ClassificationMini-Imagenet (test)--
235
Generalized Zero-Shot LearningSUN
H22
184
Skeleton-based Action RecognitionNTU RGB+D 120 (X-set)
Top-1 Accuracy30.2
184
Generalized Zero-Shot LearningAWA2
S Score39.7
165
Skeleton-based Action RecognitionNTU RGB+D 120 Cross-Subject
Top-1 Accuracy19.8
143
Generalized Zero-Shot LearningAWA1
S Score37.1
49
Action RecognitionNTU RGB+D 120 (Cross-View)
Accuracy57.92
47
Showing 10 of 89 rows
...

Other info

Follow for update