SHIFT: Steering Hidden Intermediates in Flow Transformers
About
Diffusion models have become leading approaches for high-fidelity image generation. Recent DiT-based diffusion models, in particular, achieve strong prompt adherence while producing high-quality samples. We propose SHIFT, a simple but effective and lightweight framework for concept removal in DiT diffusion models via targeted manipulation of intermediate activations at inference time, inspired by activation steering in large language models. SHIFT learns steering vectors that are dynamically applied to selected layers and timesteps to suppress unwanted visual concepts while preserving the prompt's remaining content and overall image quality. Beyond suppression, the same mechanism can shift generations into a desired \emph{style domain} or bias samples toward adding or changing target objects. We demonstrate that SHIFT provides effective and flexible control over DiT generation across diverse prompts and targets without time-consuming retraining.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Nudity Erasure | I2P (dev) | Common Count155 | 10 | |
| Style Erasure | Old artists | AccE30 | 7 | |
| Style Erasure | Modern artists | AccE60 | 7 | |
| Image Generation Evaluation | MS-COCO 5K (sampled images) | FID33.9 | 7 | |
| Concept Preservation | Mickey prompts | Concept Score (CS)26.06 | 3 | |
| Concept Preservation | Dog prompts | CS24.18 | 3 | |
| Concept Preservation | Legislator prompts | CS21.77 | 3 | |
| Concrete Object Erasure | Pikachu Concrete Object Erasure | CS25.14 | 3 | |
| Concept Preservation | Spongebob prompts | Concept Score (CS)27.35 | 3 | |
| Concept Preservation | Pikachu prompts | Concept Score (CS)26.25 | 3 |