HiLLoC: Lossless Image Compression with Hierarchical Latent Variable Models
About
We make the following striking observation: fully convolutional VAE models trained on 32x32 ImageNet can generalize well, not just to 64x64 but also to far larger photographs, with no changes to the model. We use this property, applying fully convolutional models to lossless compression, demonstrating a method to scale the VAE-based 'Bits-Back with ANS' algorithm for lossless compression to large color photographs, and achieving state of the art for compression of full size ImageNet images. We release Craystack, an open source library for convenient prototyping of lossless compression using probabilistic models, along with full implementations of all of our compression results.
James Townsend, Thomas Bird, Julius Kunze, David Barber• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Lossless Compression | CIFAR10 (test) | Bits Per Dimension (BPD)3.32 | 30 | |
| Lossless Compression | ImageNet64 (test) | BPD3.89 | 27 | |
| Lossless Compression | ImageNet32 (test) | BPD4.2 | 20 | |
| Lossless Compression | CIFAR10 | BPD3.56 | 20 | |
| Lossless Compression | SVHN (test) | BPD2.29 | 16 | |
| Lossless Compression | CelebA 32x32 (test) | BPD3.52 | 16 | |
| Lossless Image Compression | ImageNet64 | BPD3.9 | 16 | |
| Lossless Compression | ImageNet Full (test) | BPD3.25 | 14 | |
| Image Compression | ImageNet-32 | BPD4.2 | 14 | |
| Image Compression | ImageNet 64 | BPD3.9 | 13 |
Showing 10 of 13 rows