Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

h-Edit: Effective and Flexible Diffusion-Based Editing via Doob's h-Transform

About

We introduce a theoretical framework for diffusion-based image editing by formulating it as a reverse-time bridge modeling problem. This approach modifies the backward process of a pretrained diffusion model to construct a bridge that converges to an implicit distribution associated with the editing target at time 0. Building on this framework, we propose h-Edit, a novel editing method that utilizes Doob's h-transform and Langevin Monte Carlo to decompose the update of an intermediate edited sample into two components: a "reconstruction" term and an "editing" term. This decomposition provides flexibility, allowing the reconstruction term to be computed via existing inversion techniques and enabling the combination of multiple editing terms to handle complex editing tasks. To our knowledge, h-Edit is the first training-free method capable of performing simultaneous text-guided and reward-model-based editing. Extensive experiments, both quantitative and qualitative, show that h-Edit outperforms state-of-the-art baselines in terms of editing effectiveness and faithfulness. Our source code is available at https://github.com/nktoan/h-edit.

Toan Nguyen, Kien Do, Duc Kieu, Thin Nguyen• 2025

Related benchmarks

TaskDatasetResultRank
Image EditingPIE-Bench (test)
PSNR17.6958
46
Image EditingImageNetR-Fake (test)
Structural Distance44.7574
6
Showing 2 of 2 rows

Other info

Follow for update