Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers

About

We present Scalable Interpolant Transformers (SiT), a family of generative models built on the backbone of Diffusion Transformers (DiT). The interpolant framework, which allows for connecting two distributions in a more flexible way than standard diffusion models, makes possible a modular study of various design choices impacting generative models built on dynamical transport: learning in discrete or continuous time, the objective function, the interpolant that connects the distributions, and deterministic or stochastic sampling. By carefully introducing the above ingredients, SiT surpasses DiT uniformly across model sizes on the conditional ImageNet 256x256 and 512x512 benchmark using the exact same model structure, number of parameters, and GFLOPs. By exploring various diffusion coefficients, which can be tuned separately from learning, SiT achieves an FID-50K score of 2.06 and 2.62, respectively.

Nanye Ma, Mark Goldstein, Michael S. Albergo, Nicholas M. Boffi, Eric Vanden-Eijnden, Saining Xie• 2024

Related benchmarks

TaskDatasetResultRank
Class-conditional Image GenerationImageNet 256x256
Inception Score (IS)277.5
441
Image GenerationImageNet 256x256 (val)
FID2.02
307
Class-conditional Image GenerationImageNet 256x256 (train)
IS272.3
305
Class-conditional Image GenerationImageNet 256x256 (val)
FID2.06
293
Image GenerationImageNet 256x256
FID2.05
243
Image GenerationImageNet (val)
FID2.06
198
Image GenerationImageNet 512x512 (val)
FID-50K2.62
184
Class-conditional Image GenerationImageNet 256x256 (train val)
FID2.06
178
Class-conditional Image GenerationImageNet 256x256 (test)
FID2.15
167
Class-conditional Image GenerationImageNet--
132
Showing 10 of 44 rows

Other info

Follow for update