Proximal Gradient methods with Adaptive Subspace Sampling
About
Many applications in machine learning or signal processing involve nonsmooth optimization problems. This nonsmoothness brings a low-dimensional structure to the optimal solutions. In this paper, we propose a randomized proximal gradient method harnessing this underlying structure. We introduce two key components: i) a random subspace proximal gradient algorithm; ii) an identification-based sampling of the subspaces. Their interplay brings a significant performance improvement on typical learning problems in terms of dimensions explored.
Dmitry Grishchenko, Franck Iutzeler, J\'er\^ome Malick• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Precision Matrix Estimation | Banded 1 precision matrix structure | Time (s)1.209 | 24 | |
| Precision Matrix Estimation | Banded 2 | Time (s)1.376 | 24 | |
| Precision Matrix Estimation | Grid precision matrix structure | Time (s)1.49 | 24 |
Showing 3 of 3 rows