3SGen: Unified Subject, Style, and Structure-Driven Image Generation with Adaptive Task-specific Memory
About
Recent image generation approaches often address subject, style, and structure-driven conditioning in isolation, leading to feature entanglement and limited task transferability. In this paper, we introduce 3SGen, a task-aware unified framework that performs all three conditioning modes within a single model. 3SGen employs an MLLM equipped with learnable semantic queries to align text-image semantics, complemented by a VAE branch that preserves fine-grained visual details. At its core, an Adaptive Task-specific Memory (ATM) module dynamically disentangles, stores, and retrieves condition-specific priors, such as identity for subjects, textures for styles, and spatial layouts for structures, via a lightweight gating mechanism along with several scalable memory items. This design mitigates inter-task interference and naturally scales to compositional inputs. In addition, we propose 3SGen-Bench, a unified image-driven generation benchmark with standardized metrics for evaluating cross-task fidelity and controllability. Extensive experiments on our proposed 3SGen-Bench and other public benchmarks demonstrate our superior performance across diverse image-driven generation tasks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image-driven Generation | 3SGen-Bench | Subject Fidelity Score8.41 | 6 | |
| Style-driven Generation | Multi-task Image-driven Generation Evaluation Set | CSD52 | 6 | |
| Subject-driven generation | Multi-task Image-driven Generation Evaluation Set | CLIP-I0.638 | 6 | |
| Structure-driven Generation | Multi-task Image-driven Generation Evaluation Set | Struc-Sim35.71 | 4 |