Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Class Distribution Shifts in Zero-Shot Learning: Learning Robust Representations

About

Zero-shot learning methods typically assume that the new, unseen classes encountered during deployment come from the same distribution as the the classes in the training set. However, real-world scenarios often involve class distribution shifts (e.g., in age or gender for person identification), posing challenges for zero-shot classifiers that rely on learned representations from training classes. In this work, we propose and analyze a model that assumes that the attribute responsible for the shift is unknown in advance. We show that in this setting, standard training may lead to non-robust representations. To mitigate this, we develop an algorithm for learning robust representations in which (a) synthetic data environments are constructed via hierarchical sampling, and (b) environment balancing penalization, inspired by out-of-distribution problems, is applied. We show that our algorithm improves generalization to diverse class distributions in both simulations and experiments on real-world datasets.

Yuli Slavutsky, Yuval Benjamini• 2023

Related benchmarks

TaskDatasetResultRank
VerificationSimulation In Distribution (test)
AUC95
18
VerificationSimulation Distribution Shift (test)
AUC88.1
18
Zero-shot ClassificationCELEBA distribution shift
AUC69.7
5
Zero-shot ClassificationETHEC distribution shift
AUC83.8
5
Zero-shot ClassificationCELEBA (in-distribution)
AUC83.6
5
Zero-shot ClassificationETHEC (in-distribution)
AUC87.2
5
Showing 6 of 6 rows

Other info

Code

Follow for update