How To Make the Gradients Small Stochastically

01/08/2018
by   Zeyuan Allen-Zhu, et al.
0

In convex stochastic optimization, convergence rates in terms of minimizing the objective have been well-established. However, in terms of making the gradients small, the best known convergence rate was O(ε^-8/3) and it was left open how to improve it. In this paper, we improve this rate to Õ(ε^-2), which is optimal up to log factors.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro