Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

PersonalQ: Select, Quantize, and Serve Personalized Diffusion Models for Efficient Inference

About

Personalized text-to-image generation lets users fine-tune diffusion models into repositories of concept-specific checkpoints, but serving these repositories efficiently is difficult for two reasons: natural-language requests are often ambiguous and can be misrouted to visually similar checkpoints, and standard post-training quantization can distort the fragile representations that encode personalized concepts. We present PersonalQ, a unified framework that connects checkpoint selection and quantization through a shared signal -- the checkpoint's trigger token. Check-in performs intent-aligned selection by combining intent-aware hybrid retrieval with LLM-based reranking over checkpoint context and asks a brief clarification question only when multiple intents remain plausible; it then rewrites the prompt by inserting the selected checkpoint's canonical trigger. Complementing this, Trigger-Aware Quantization (TAQ) applies trigger-aware mixed precision in cross-attention, preserving trigger-conditioned key/value rows (and their attention weights) while aggressively quantizing the remaining pathways for memory-efficient inference. Experiments show that PersonalQ improves intent alignment over retrieval and reranking baselines, while TAQ consistently offers a stronger compression-quality trade-off than prior diffusion PTQ methods, enabling scalable serving of personalized checkpoints without sacrificing fidelity.

Qirui Wang, Qi Guo, Yiding Sun, Junkai Yang, Dongxu Zhang, Shanmin Pang, Qing Guo• 2026

Related benchmarks

TaskDatasetResultRank
Personalized Text-to-Image GenerationMS-COCO
FID11.03
18
Personalized Text-to-Image GenerationPartiPrompts
FID10.49
18
Checkpoint SelectionREPO-PROMPTS, MS-COCO, and PartiPrompts
Intent Score4.42
4
Showing 3 of 3 rows

Other info

Follow for update