Leveraging sparse and shared feature activations for disentangled representation learning
About
Recovering the latent factors of variation of high dimensional data has so far focused on simple synthetic settings. Mostly building on unsupervised and weakly-supervised objectives, prior work missed out on the positive implications for representation learning on real world data. In this work, we propose to leverage knowledge extracted from a diversified set of supervised tasks to learn a common disentangled representation. Assuming each supervised task only depends on an unknown subset of the factors of variation, we disentangle the feature space of a supervised multi-task model, with features activating sparsely across different tasks and information being shared as appropriate. Importantly, we never directly observe the factors of variations but establish that access to multiple tasks is sufficient for identifiability under sufficiency and minimality assumptions. We validate our approach on six real world distribution shift benchmarks, and different data modalities (images, text), demonstrating how disentangled representations can be transferred to real settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Domain Generalization | VLCS | Accuracy77.3 | 238 | |
| Domain Generalization | PACS | -- | 221 | |
| Domain Generalization | OfficeHome | Accuracy82 | 182 | |
| Few-shot Image Classification | miniImageNet (test) | Accuracy76.6 | 111 | |
| Classification | CivilComments (test) | Worst-case Accuracy75.45 | 47 | |
| Few-shot Image Classification | CIFAR FS (test) | Accuracy86.9 | 46 | |
| Domain Generalization | VLCS DomainBed (test) | Average OOD Accuracy78.4 | 27 | |
| Domain Generalization | OfficeHome DomainBed (OOD) | Avg OOD Accuracy70.9 | 16 | |
| Domain Generalization | PACS OOD (test) | Average Accuracy90.4 | 13 | |
| Image Classification | Camelyon17-WILDS out-of-distribution (val) | Accuracy89.9 | 9 |