Parameter-free version of Adaptive Gradient Methods for Strongly-Convex Functions

06/11/2023
by   Deepak Gouda, et al.
0

The optimal learning rate for adaptive gradient methods applied to λ-strongly convex functions relies on the parameters λ and learning rate η. In this paper, we adapt a universal algorithm along the lines of Metagrad, to get rid of this dependence on λ and η. The main idea is to concurrently run multiple experts and combine their predictions to a master algorithm. This master enjoys O(d log T) regret bounds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/15/2019

Adaptivity and Optimality: A Universal Algorithm for Online Convex Optimization

In this paper, we study adaptive online convex optimization, and aim to ...
research
06/26/2019

Dual Adaptivity: A Universal Algorithm for Minimizing the Adaptive Regret of Convex Functions

To deal with changing environments, a new performance measure---adaptive...
research
09/12/2023

ELRA: Exponential learning rate adaption gradient descent optimization method

We present a novel, fast (exponential rate adaption), ab initio (hyper-p...
research
03/27/2023

Learning Rate Schedules in the Presence of Distribution Shift

We design learning rate schedules that minimize regret for SGD-based onl...
research
12/19/2016

Corralling a Band of Bandit Algorithms

We study the problem of combining multiple bandit algorithms (that is, o...
research
09/05/2018

Anytime Hedge achieves optimal regret in the stochastic regime

This paper is about a surprising fact: we prove that the anytime Hedge a...
research
08/02/2018

Online Aggregation of Unbounded Losses Using Shifting Experts with Confidence

We develop the setting of sequential prediction based on shifting expert...

Please sign up or login with your details

Forgot password? Click here to reset