Fine-Tuning Discrete Diffusion Models with Policy Gradient Methods
About
Discrete diffusion models have recently gained significant attention due to their ability to process complex discrete structures for language modeling. However, fine-tuning these models with policy gradient methods, as is commonly done in Reinforcement Learning from Human Feedback (RLHF), remains a challenging task. We propose an efficient, broadly applicable, and theoretically justified policy gradient algorithm, called Score Entropy Policy Optimization (\SEPO), for fine-tuning discrete diffusion models over non-differentiable rewards. Our numerical experiments across several discrete generative tasks demonstrate the scalability and efficiency of our method. Our code is available at https://github.com/ozekri/SEPO.
Oussama Zekri, Nicolas Boull\'e• 2025
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| DNA enhancer design | HepG2 (test) | Pred. Activity7.55 | 6 |
Showing 1 of 1 rows