Pinet: Optimizing hard-constrained neural networks with orthogonal projection layers
About
We introduce an output layer for neural networks that ensures satisfaction of convex constraints. Our approach, $\Pi$net, leverages operator splitting for rapid and reliable projections in the forward pass, and the implicit function theorem for backpropagation. We deploy $\Pi$net as a feasible-by-design optimization proxy for parametric constrained optimization problems and obtain modest-accuracy solutions faster than traditional solvers when solving a single problem, and significantly faster for a batch of problems. We surpass state-of-the-art learning approaches by orders of magnitude in terms of training time, solution quality, and robustness to hyperparameter tuning, while maintaining similar inference times. Finally, we tackle multi-vehicle motion planning with non-convex trajectory preferences and provide $\Pi$net as a GPU-ready package implemented in JAX.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Inference Time Estimation | Convex large | Median Time6.3 | 10 | |
| Inference Time Estimation | Convex small | Median Latency0.0055 | 10 | |
| Optimization Solver Inference Time | Non-convex large (test) | Inference Time (Median)0.0063 | 8 | |
| Second-order cone programming | Large second-order cone programs (test) | Median Runtime (s)0.0122 | 8 | |
| Optimization Solver Inference Time | Non-convex small (test) | Median Latency0.0056 | 8 | |
| Second-order cone programming | Small second-order cone programs (test) | Median Runtime (s)0.0091 | 8 | |
| Motion Planning | Non-convex small motion planning problems | Median Latency (s)0.0056 | 6 | |
| Motion Planning | Non-convex large motion planning problems | Median Latency0.0063 | 6 | |
| Constrained Optimization | Small Non-Convex benchmark (test) | RS0.0035 | 3 |