Domain Generalization by Mutual-Information Regularization with Pre-trained Models
About
Domain generalization (DG) aims to learn a generalized model to an unseen target domain using only limited source domains. Previous attempts to DG fail to learn domain-invariant representations only from the source domains due to the significant domain shifts between training and test domains. Instead, we re-formulate the DG objective using mutual information with the oracle model, a model generalized to any possible domain. We derive a tractable variational lower bound via approximating the oracle model by a pre-trained model, called Mutual Information Regularization with Oracle (MIRO). Our extensive experiments show that MIRO significantly improves the out-of-distribution performance. Furthermore, our scaling experiments show that the larger the scale of the pre-trained model, the greater the performance improvement of MIRO. Source code is available at https://github.com/kakaobrain/miro.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Domain Generalization | VLCS | Accuracy79.1 | 238 | |
| Domain Generalization | PACS | Accuracy (Art)89.3 | 221 | |
| Domain Generalization | OfficeHome | Accuracy70.7 | 182 | |
| Image Classification | DomainNet | Accuracy (ClipArt)74.9 | 161 | |
| Image Classification | OfficeHome | Average Accuracy70.5 | 131 | |
| Domain Generalization | DomainBed | Average Accuracy77.3 | 127 | |
| Domain Generalization | DomainNet | Accuracy44.3 | 113 | |
| Domain Generalization | DomainBed (test) | VLCS Accuracy79.9 | 110 | |
| Image Classification | PACS | Accuracy85.4 | 100 | |
| Domain Generalization | TerraIncognita | Accuracy50.4 | 81 |