Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Functional Gradient Flows for Constrained Sampling

About

Recently, through a unified gradient flow perspective of Markov chain Monte Carlo (MCMC) and variational inference (VI), particle-based variational inference methods (ParVIs) have been proposed that tend to combine the best of both worlds. While typical ParVIs such as Stein Variational Gradient Descent (SVGD) approximate the gradient flow within a reproducing kernel Hilbert space (RKHS), many attempts have been made recently to replace RKHS with more expressive function spaces, such as neural networks. While successful, these methods are mainly designed for sampling from unconstrained domains. In this paper, we offer a general solution to constrained sampling by introducing a boundary condition for the gradient flow which would confine the particles within the specific domain. This allows us to propose a new functional gradient ParVI method for constrained sampling, called constrained functional gradient flow (CFG), with provable continuous-time convergence in total variation (TV). We also present novel numerical strategies to handle the boundary integral term arising from the domain constraints. Our theory and experiments demonstrate the effectiveness of the proposed framework.

Shiyue Zhang, Longlin Yu, Ziheng Cheng, Cheng Zhang• 2024

Related benchmarks

TaskDatasetResultRank
Monotonic Bayesian Neural Network ClassificationCOMPAS (test)
OOD Ratio0.00e+0
24
Monotonic Bayesian Neural Network RegressionBlog Feedback (test)
Ratio Out (%)0.00e+0
4
Target distribution approximationCardioid
Wasserstein-2 Distance (Sinkhorn)0.1141
2
Target distribution approximationDouble-moon
Wasserstein-2 Distance (Sinkhorn)0.166
2
Target distribution approximationRing
Wasserstein-2 Distance (Sinkhorn)0.1087
2
Showing 5 of 5 rows

Other info

Code

Follow for update