Provably Faster Algorithms for Bilevel Optimization

06/08/2021
by   Junjie Yang, et al.
0

Bilevel optimization has been widely applied in many important machine learning applications such as hyperparameter optimization and meta-learning. Recently, several momentum-based algorithms have been proposed to solve bilevel optimization problems faster. However, those momentum-based algorithms do not achieve provably better computational complexity than 𝒪(ϵ^-2) of the SGD-based algorithm. In this paper, we propose two new algorithms for bilevel optimization, where the first algorithm adopts momentum-based recursive iterations, and the second algorithm adopts recursive gradient estimations in nested loops to decrease the variance. We show that both algorithms achieve the complexity of 𝒪(ϵ^-1.5), which outperforms all existing algorithms by the order of magnitude. Our experiments validate our theoretical results and demonstrate the superior empirical performance of our algorithms in hyperparameter applications. Our codes for MRBO, VRBO and other benchmarks are available online^1.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/03/2022

Faster Adaptive Momentum-Based Federated Methods for Distributed Composition Optimization

Composition optimization recently appears in many machine learning appli...
research
07/25/2019

DEAM: Accumulated Momentum with Discriminative Weight for Stochastic Optimization

Optimization algorithms with momentum, e.g., Nesterov Accelerated Gradie...
research
04/28/2022

On the Convergence of Momentum-Based Algorithms for Federated Stochastic Bilevel Optimization Problems

In this paper, we studied the federated stochastic bilevel optimization ...
research
11/02/2022

Fast Adaptive Federated Bilevel Optimization

Bilevel optimization is a popular hierarchical model in machine learning...
research
05/03/2022

Local Stochastic Bilevel Optimization with Momentum-Based Variance Reduction

Bilevel Optimization has witnessed notable progress recently with new em...
research
02/07/2019

Momentum Schemes with Stochastic Variance Reduction for Nonconvex Composite Optimization

Two new stochastic variance-reduced algorithms named SARAH and SPIDER ha...
research
05/25/2020

Triangularized Orthogonalization-free Method for Solving Extreme Eigenvalue Problems

A novel orthogonalization-free method together with two specific algorit...

Please sign up or login with your details

Forgot password? Click here to reset