Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On Graduated Optimization for Stochastic Non-Convex Problems

About

The graduated optimization approach, also known as the continuation method, is a popular heuristic to solving non-convex problems that has received renewed interest over the last decade. Despite its popularity, very little is known in terms of theoretical convergence analysis. In this paper we describe a new first-order algorithm based on graduated optimiza- tion and analyze its performance. We characterize a parameterized family of non- convex functions for which this algorithm provably converges to a global optimum. In particular, we prove that the algorithm converges to an {\epsilon}-approximate solution within O(1/\epsilon^2) gradient-based steps. We extend our algorithm and analysis to the setting of stochastic non-convex optimization with noisy gradient feedback, attaining the same convergence rate. Additionally, we discuss the setting of zero-order optimization, and devise a a variant of our algorithm which converges at rate of O(d^2/\epsilon^4).

Elad Hazan, Kfir Y. Levy, Shai Shalev-Shwartz• 2015

Related benchmarks

TaskDatasetResultRank
Adversarial AttackMNIST 100 images
Success Rate84
5
Black-box Adversarial AttackCIFAR-10 (test)
Success Rate65
5
Showing 2 of 2 rows

Other info

Follow for update