Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VGSE: Visually-Grounded Semantic Embeddings for Zero-Shot Learning

About

Human-annotated attributes serve as powerful semantic embeddings in zero-shot learning. However, their annotation process is labor-intensive and needs expert supervision. Current unsupervised semantic embeddings, i.e., word embeddings, enable knowledge transfer between classes. However, word embeddings do not always reflect visual similarities and result in inferior zero-shot performance. We propose to discover semantic embeddings containing discriminative visual properties for zero-shot learning, without requiring any human annotation. Our model visually divides a set of images from seen classes into clusters of local image regions according to their visual similarity, and further imposes their class discrimination and semantic relatedness. To associate these clusters with previously unseen classes, we use external knowledge, e.g., word embeddings and propose a novel class relation discovery module. Through quantitative and qualitative evaluation, we demonstrate that our model discovers semantic embeddings that model the visual properties of both seen and unseen classes. Furthermore, we demonstrate on three benchmarks that our visually-grounded semantic embeddings further improve performance over word embeddings across various ZSL models by a large margin.

Wenjia Xu, Yongqin Xian, Jiuniu Wang, Bernt Schiele, Zeynep Akata• 2022

Related benchmarks

TaskDatasetResultRank
Generalized Zero-Shot LearningCUB
H Score31.5
250
Generalized Zero-Shot LearningSUN
H29.8
184
Zero-shot LearningSUN (unseen)
Top-1 Accuracy (%)41.1
50
Zero-shot LearningCUB (unseen)
Top-1 Accuracy35
49
Zero-shot LearningAWA2 (unseen)
Top-1 Acc64
37
Generalized Zero-Shot LearningAWA2 (seen unseen)
U Score51.2
10
Showing 6 of 6 rows

Other info

Code

Follow for update