We present a simple restart scheme for speeding-up first order methods for convex optimization problems. Restarts are made based only on achieving specified decreases in objective values, the specified amounts being the same for all optimization problems. Unlike existing restart schemes, the scheme makes no attempt to learn parameter values characterizing the structure of an optimization problem, nor does it require any special information that would not be available in practice (unless the first-order method chosen to be employed in the scheme itself requires special information). As immediate corollaries to the main theorems, we show that when some well-known first order methods are employed in the scheme, the resulting complexity bounds are nearly optimal for particular – yet quite general – classes of problems. We close with discussion of ongoing extensions to the work.
Tuesday, April 24, 2018 - 4:00pm