Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Prompt Optimization Via Diffusion Language Models

About

We propose a diffusion-based framework for prompt optimization that leverages Diffusion Language Models (DLMs) to iteratively refine system prompts through masked denoising. By conditioning on interaction traces, including user queries, model responses, and optional feedback, our method enables flexible, span-level prompt updates without requiring gradient access or modifying the downstream language model. Across diverse benchmarks (e.g., $\tau$-bench, SST-2, SST-5), DLM-optimized prompts consistently improve the performance of a frozen target LLM (e.g., GPT-4o-mini). We further show that moderate diffusion step counts provide the best balance between refinement quality and stability. These results highlight diffusion-based prompt optimization as a general, model-agnostic, and scalable approach for enhancing LLM performance through iterative prompt refinement.

Shiyu Wang, Haolin Chen, Liangwei Yang, Jielin Qiu, Rithesh Murthy, Ming Zhu, Zixiang Chen, Silvio Savarese, Caiming Xiong, Shelby Heinecke, Huan Wang• 2026

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI
Accuracy93
174
Function CallingTau-bench airline
Success Rate50
5
Function CallingTau-bench retail
Success Rate46
5
Sentiment AnalysisSST-2
Success Rate97
5
Sentiment AnalysisSST-5
Success Rate (SST-5)67
5
Semantic EquivalenceMRPC
Success Rate69
5
Showing 6 of 6 rows

Other info

Follow for update