Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

FoundIR-v2: Optimizing Pre-Training Data Mixtures for Image Restoration Foundation Model

About

Recent studies have witnessed significant advances in image restoration foundation models driven by improvements in the scale and quality of pre-training data. In this work, we find that the data mixture proportions from different restoration tasks are also a critical factor directly determining the overall performance of all-in-one image restoration models. To this end, we propose a high-capacity diffusion-based image restoration foundation model, FoundIR-v2, which adopts a data equilibrium scheduling paradigm to dynamically optimize the proportions of mixed training datasets from different tasks. By leveraging the data mixing law, our method ensures a balanced dataset composition, enabling the model to achieve consistent generalization and comprehensive performance across diverse tasks. Furthermore, we introduce an effective Mixture-of-Experts (MoE)-driven scheduler into generative pre-training to flexibly allocate task-adaptive diffusion priors for each restoration task, accounting for the distinct degradation forms and levels exhibited by different tasks. Extensive experiments demonstrate that our method can address over 50 sub-tasks across a broader scope of real-world scenarios and achieves favorable performance against state-of-the-art approaches.

Xiang Chen, Jinshan Pan, Jiangxin Dong, Jian Yang, Jinhui Tang• 2025

Related benchmarks

TaskDatasetResultRank
DesnowingWeatherBench
PSNR23.15
17
Defocus DeblurringLSD
PSNR20.78
10
DehazingDense-Haze
PSNR15.29
10
DerainingHQ-NightRain
PSNR18.57
10
Joint Denoising and EnhancementFoundIR-L+N
PSNR19.72
10
Motion Deblurring4KRD
PSNR26.64
10
Non-Homogeneous DehazingNH-HAZE
PSNR17
10
Raindrop RemovalUAV-Rain1k
PSNR17.9
10
Super-ResolutionRealPhoto60
MUSIQ72.36
10
UHD EnhancementUHD-LL
PSNR20.08
10
Showing 10 of 16 rows

Other info

Follow for update