Intervening to Learn and Compose Causally Disentangled Representations
About
In designing generative models, it is commonly believed that in order to learn useful latent structure, we face a fundamental tension between expressivity and structure. In this paper we challenge this view by proposing a new approach to training arbitrarily expressive generative models that simultaneously learn causally disentangled concepts. This is accomplished by adding a simple context module to an arbitrarily complex black-box model, which learns to process concept information by implicitly inverting linear representations from the model's encoder. Inspired by the notion of intervention in a causal model, our module selectively modifies its architecture during training, allowing it to learn a compact joint model over different contexts. We show how adding this module leads to causally disentangled representations that can be composed for out-of-distribution generation on both real and simulated data. The resulting models can be trained end-to-end or fine-tuned from pre-trained models. To further validate our proposed approach, we prove a new identifiability result that extends existing work on identifying structured representations.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Reconstruction | MNIST 1.0 (test) | Recon Error0.135 | 9 | |
| Concept Composition | quad independent 1.0 (test) | Composition Score0.306 | 9 | |
| Concept Learning | quad independent 1.0 (test) | Concept Score20.2 | 9 | |
| Reconstruction | quad independent 1.0 (test) | Recon0.84 | 9 | |
| Reconstruction | quad dependent 1.0 (test) | Recon Score0.889 | 9 | |
| Concept Composition | 3DIdent 1.0 (test) | Compo6.8 | 9 | |
| Reconstruction | 3DIdent 1.0 (test) | Reconstruction Score0.609 | 9 | |
| Concept Composition | quad dependent 1.0 (test) | Compo Score20.3 | 9 | |
| Concept Learning | quad dependent 1.0 (test) | Concept Score16.6 | 9 | |
| Concept Learning | MNIST 1.0 (test) | Concept9.9 | 9 |