How To Make the Gradients Small Stochastically

01/08/2018
by   Zeyuan Allen-Zhu, et al.
0

In convex stochastic optimization, convergence rates in terms of minimizing the objective have been well-established. However, in terms of making the gradients small, the best known convergence rate was O(ε^-8/3) and it was left open how to improve it. In this paper, we improve this rate to Õ(ε^-2), which is optimal up to log factors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/12/2017

Convergence Rates for Deterministic and Stochastic Subgradient Methods Without Lipschitz Continuity

We generalize the classic convergence rate theory for subgradient method...
research
09/11/2019

The Error-Feedback Framework: Better Rates for SGD with Delayed Gradients and Compressed Communication

We analyze (stochastic) gradient descent (SGD) with delayed updates on s...
research
06/11/2018

Convergence Rates for Projective Splitting

Projective splitting is a family of methods for solving inclusions invol...
research
07/18/2012

Stochastic optimization and sparse statistical recovery: An optimal algorithm for high dimensions

We develop and analyze stochastic optimization algorithms for problems i...
research
09/11/2021

Fundamental limits of over-the-air optimization: Are analog schemes optimal?

We consider over-the-air convex optimization on a d-dimensional space wh...
research
03/06/2023

Convergence Rates for Non-Log-Concave Sampling and Log-Partition Estimation

Sampling from Gibbs distributions p(x) ∝exp(-V(x)/ε) and computing their...
research
08/22/2019

Finite Precision Stochastic Optimization -- Accounting for the Bias

We consider first order stochastic optimization where the oracle must qu...

Please sign up or login with your details

Forgot password? Click here to reset