Stein Variational Model Predictive Control
About
Decision making under uncertainty is critical to real-world, autonomous systems. Model Predictive Control (MPC) methods have demonstrated favorable performance in practice, but remain limited when dealing with complex probability distributions. In this paper, we propose a generalization of MPC that represents a multitude of solutions as posterior distributions. By casting MPC as a Bayesian inference problem, we employ variational methods for posterior computation, naturally encoding the complexity and multi-modality of the decision making problem. We present a Stein variational gradient descent method to estimate the posterior directly over control parameters, given a cost function and observed state trajectories. We show that this framework leads to successful planning in challenging, non-convex optimal control problems.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Control | Cart Inverted Pendulum System | MSE (x)1.26 | 8 | |
| Navigation | 2D Navigation | Collision Rate0.00e+0 | 7 | |
| Reach | Kinova manipulation suite | SR@100%100 | 7 | |
| Reach | Reach Obstacles | Collision Rate (%)0.0287 | 7 | |
| Arm Pushing Task | Franka Panda arm block pushing simulation | Mean End Distance Error (mm)19.36 | 7 | |
| Reach-avoid navigation | Single-agent reach-avoid 2D point mass model | Runtime (ms)178.4 | 6 | |
| Reach (Obstacles) | Kinova manipulation suite | SR @ 100%20 | 5 | |
| Point-to-point Navigation | Unknown Cluttered Environments | Success Rate64 | 4 | |
| Trajectory Optimization | 8x8m Open Environment Goal (6, 0) | Success Rate100 | 4 | |
| Trajectory Optimization | 8x8m Cluttered Environment Goal (6, 0) | Success Rate (SR)100 | 4 |