Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Lossy Compression with Gaussian Diffusion

About

We consider a novel lossy compression approach based on unconditional diffusion generative models, which we call DiffC. Unlike modern compression schemes which rely on transform coding and quantization to restrict the transmitted information, DiffC relies on the efficient communication of pixels corrupted by Gaussian noise. We implement a proof of concept and find that it works surprisingly well despite the lack of an encoder transform, outperforming the state-of-the-art generative compression method HiFiC on ImageNet 64x64. DiffC only uses a single model to encode and denoise corrupted pixels at arbitrary bitrates. The approach further provides support for progressive coding, that is, decoding from partial bit streams. We perform a rate-distortion analysis to gain a deeper understanding of its performance, providing analytical results for multivariate Gaussian data as well as theoretic bounds for general distributions. Furthermore, we prove that a flow-based reconstruction achieves a 3 dB gain over ancestral sampling at high bitrates.

Lucas Theis, Tim Salimans, Matthew D. Hoffman, Fabian Mentzer• 2022

Related benchmarks

TaskDatasetResultRank
Image CompressionDIV2K 512
BD-PSNR-0.27
90
Image CompressionKodak24 512
PSNR25.81
76
Image CompressionKodak24 512x512 (test)
BD-PSNR-0.98
13
Image CompressionKodak24 (test)
PSNR20.92
8
Robust Image CompressionKodak24
PSNR14.85
8
Robust Image CompressionDIV2K
PSNR13.48
8
Image CompressionDIV2K (val)
PSNR18.62
8
Showing 7 of 7 rows

Other info

Follow for update