Directional Textual Inversion for Personalized Text-to-Image Generation
About
Textual Inversion (TI) is an efficient approach to text-to-image personalization but often fails on complex prompts. We trace these failures to embedding norm inflation: learned tokens drift to out-of-distribution magnitudes, degrading prompt conditioning in pre-norm Transformers. Empirically, we show semantics are primarily encoded by direction in CLIP token space, while inflated norms harm contextualization; theoretically, we analyze how large magnitudes attenuate positional information and hinder residual updates in pre-norm blocks. We propose Directional Textual Inversion (DTI), which fixes the embedding magnitude to an in-distribution scale and optimizes only direction on the unit hypersphere via Riemannian SGD. We cast direction learning as MAP with a von Mises-Fisher prior, yielding a constant-direction prior gradient that is simple and efficient to incorporate. Across personalization tasks, DTI improves text fidelity over TI and TI-variants while maintaining subject similarity. Crucially, DTI's hyperspherical parameterization enables smooth, semantically coherent interpolation between learned concepts (slerp), a capability that is absent in standard TI. Our findings suggest that direction-only optimization is a robust and scalable path for prompt-faithful personalization.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Personalized Text-to-Image Generation | SD 1.5 | Image Fidelity Score0.418 | 4 | |
| Subject-driven image generation | SANA 1.5-4.8B (Evaluation Set) | Image Alignment45.2 | 4 | |
| Personalized Text-to-Image Generation | SD base 2.1 | Image Fidelity0.469 | 4 | |
| Subject-driven image generation | SANA 1.5-1.6B Evaluation Set | Image Fidelity0.479 | 4 | |
| Subject-driven image generation | SDXL Evaluation Set | Image Score0.45 | 4 | |
| Text-to-Image Generation | Amazon Mechanical Turk User Study (test) | Image Fidelity43.45 | 3 |