Automatically Bounding the Taylor Remainder Series: Tighter Bounds and New Applications

12/22/2022
by   Matthew Streeter, et al.
0

We present a new algorithm for automatically bounding the Taylor remainder series. In the special case of a scalar function f: ℝ→ℝ, our algorithm takes as input a reference point x_0, trust region [a, b], and integer k ≥ 1, and returns an interval I such that f(x) - ∑_i=0^k-11/i! f^(i)(x_0) (x - x_0)^i ∈ I (x - x_0)^k for all x ∈ [a, b]. As in automatic differentiation, the function f is provided to the algorithm in symbolic form, and must be composed of known atomic functions. At a high level, our algorithm has two steps. First, for a variety of commonly-used elementary functions (e.g., exp, log), we derive sharp polynomial upper and lower bounds on the Taylor remainder series. We then recursively combine the bounds for the elementary functions using an interval arithmetic variant of Taylor-mode automatic differentiation. Our algorithm can make efficient use of machine learning hardware accelerators, and we provide an open source implementation in JAX. We then turn our attention to applications. Most notably, we use our new machinery to create the first universal majorization-minimization optimization algorithms: algorithms that iteratively minimize an arbitrary loss using a majorizer that is derived automatically, rather than by hand. Applied to machine learning, this leads to architecture-specific optimizers for training deep networks that converge from any starting point, without hyperparameter tuning. Our experiments show that for some optimization problems, these hyperparameter-free optimizers outperform tuned versions of gradient descent, Adam, and AdaGrad. We also show that our automatically-derived bounds can be used for verified global optimization and numerical integration, and to prove sharper versions of Jensen's inequality.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2023

Universal Majorization-Minimization Algorithms

Majorization-minimization (MM) is a family of optimization methods that ...
research
08/01/2023

Sharp Taylor Polynomial Enclosures in One Dimension

It is often useful to have polynomial upper or lower bounds on a one-dim...
research
04/09/2020

Automatic Differentiation in ROOT

In mathematics and computer algebra, automatic differentiation (AD) is a...
research
05/31/2021

Efficient and Modular Implicit Differentiation

Automatic differentiation (autodiff) has revolutionized machine learning...
research
04/20/2018

Autotune: A Derivative-free Optimization Framework for Hyperparameter Tuning

Machine learning applications often require hyperparameter tuning. The h...
research
06/29/2020

On the Iteration Complexity of Hypergradient Computation

We study a general class of bilevel problems, consisting in the minimiza...

Please sign up or login with your details

Forgot password? Click here to reset