Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Invertible Residual Networks

About

We show that standard ResNet architectures can be made invertible, allowing the same model to be used for classification, density estimation, and generation. Typically, enforcing invertibility requires partitioning dimensions or restricting network architectures. In contrast, our approach only requires adding a simple normalization step during training, already available in standard frameworks. Invertible ResNets define a generative model which can be trained by maximum likelihood on unlabeled data. To compute likelihoods, we introduce a tractable approximation to the Jacobian log-determinant of a residual block. Our empirical evaluation shows that invertible ResNets perform competitively with both state-of-the-art image classifiers and flow-based generative models, something that has not been previously achieved with a single architecture.

Jens Behrmann, Will Grathwohl, Ricky T. Q. Chen, David Duvenaud, J\"orn-Henrik Jacobsen• 2018

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID65.01
471
Unconditional Image GenerationCIFAR-10 (test)--
216
Density EstimationCIFAR-10 (test)
Bits/dim3.45
134
Density EstimationMNIST (test)
NLL (bits/dim)1.05
56
Generative ModelingCIFAR-10
BPD3.45
46
Density EstimationCIFAR-10
bpd3.45
40
SamplingCIFAR-10
Sampling Time (s)99.41
39
Likelihood EstimationCIFAR-10 (test)
NLL (BPD)3.45
24
Density EstimationMNIST
bpd1.06
12
Generative ModelingMNIST
BPD1.05
10
Showing 10 of 17 rows

Other info

Code

Follow for update