Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Model-Based Domain Generalization

About

Despite remarkable success in a variety of applications, it is well-known that deep learning can fail catastrophically when presented with out-of-distribution data. Toward addressing this challenge, we consider the domain generalization problem, wherein predictors are trained using data drawn from a family of related training domains and then evaluated on a distinct and unseen test domain. We show that under a natural model of data generation and a concomitant invariance condition, the domain generalization problem is equivalent to an infinite-dimensional constrained statistical learning problem; this problem forms the basis of our approach, which we call Model-Based Domain Generalization. Due to the inherent challenges in solving constrained optimization problems in deep learning, we exploit nonconvex duality theory to develop unconstrained relaxations of this statistical problem with tight bounds on the duality gap. Based on this theoretical motivation, we propose a novel domain generalization algorithm with convergence guarantees. In our experiments, we report improvements of up to 30 percentage points over state-of-the-art domain generalization baselines on several benchmarks including ColoredMNIST, Camelyon17-WILDS, FMoW-WILDS, and PACS.

Alexander Robey, George J. Pappas, Hamed Hassani• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationPACS (test)
Average Accuracy85.6
254
Multi-class classificationVLCS
Acc (Caltech)98.3
139
Image ClassificationColoredMNIST (test)
Accuracy (+90% Confidence)0.737
18
Medical Image ClassificationCamelyon17 OOD WILDS (test)
Average Accuracy93.3
13
Disease ClassificationCAMELYON17-WILDS OOD (val)
Accuracy88.1
6
Showing 5 of 5 rows

Other info

Follow for update