Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Optimal Representations for Covariate Shift

About

Machine learning systems often experience a distribution shift between training and testing. In this paper, we introduce a simple variational objective whose optima are exactly the set of all representations on which risk minimizers are guaranteed to be robust to any distribution shift that preserves the Bayes predictor, e.g., covariate shifts. Our objective has two components. First, a representation must remain discriminative for the task, i.e., some predictor must be able to simultaneously minimize the source and target risk. Second, the representation's marginal support needs to be the same across source and target. We make this practical by designing self-supervised objectives that only use unlabelled data and augmentations to train robust representations. Our objectives give insights into the robustness of CLIP, and further improve CLIP's representations to achieve SOTA results on DomainBed.

Yangjun Ruan, Yann Dubois, Chris J. Maddison• 2021

Related benchmarks

TaskDatasetResultRank
Domain GeneralizationPACS (test)
Average Accuracy64.8
225
Domain GeneralizationDomainBed v1.0 (test)
Average Accuracy82.3
71
Image ClassificationOfficeHome DomainBed suite (test)
Accuracy61
45
Domain GeneralizationDomainNet DomainBed (test)
Clipart Accuracy46.1
37
Image ClassificationDomainBed
PACS Accuracy81.9
33
Domain GeneralizationPACS DomainBed (test)
Accuracy97.2
29
Domain GeneralizationOfficeHome DomainBed (test)
Accuracy86.3
29
Domain GeneralizationPACS, VLCS, OfficeHome, and DomainNet (test)
PACS Accuracy90
28
Domain GeneralizationVLCS DomainBed (test)
Average OOD Accuracy76.1
27
Domain GeneralizationTerraIncognita DomainBed (test)
Accuracy76.5
26
Showing 10 of 16 rows

Other info

Code

Follow for update