Deterministic equivalent and error universality of deep random features learning

02/01/2023
by   Dominik Schröder, et al.
7

This manuscript considers the problem of learning a random Gaussian network function using a fully connected network with frozen intermediate layers and trainable readout layer. This problem can be seen as a natural generalization of the widely studied random features model to deeper architectures. First, we prove Gaussian universality of the test error in a ridge regression setting where the learner and target networks share the same intermediate layers, and provide a sharp asymptotic formula for it. Establishing this result requires proving a deterministic equivalent for traces of the deep random features sample covariance matrices which can be of independent interest. Second, we conjecture the asymptotic Gaussian universality of the test error in the more general setting of arbitrary convex losses and generic learner/target architectures. We provide extensive numerical evidence for this conjecture, which requires the derivation of closed-form expressions for the layer-wise post-activation population covariances. In light of our results, we investigate the interplay between architecture design and implicit regularization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2023

Precise Asymptotic Analysis of Deep Random Feature Models

We provide exact asymptotic expressions for the performance of regressio...
research
02/01/2023

Optimal Learning of Deep Random Networks of Extensive-width

We consider the problem of learning a target function corresponding to a...
research
05/17/2022

Sharp asymptotics on the compression of two-layer neural networks

In this paper, we study the compression of a target two-layer neural net...
research
11/23/2022

Quantitative deterministic equivalent of sample covariance matrices with a general dependence structure

We study sample covariance matrices arising from rectangular random matr...
research
09/16/2020

Universality Laws for High-Dimensional Learning with Random Features

We prove a universality theorem for learning with random features. Our r...
research
09/03/2022

The Gaussian product inequality conjecture for multinomial covariances

In this paper, we find an equivalent combinatorial condition only involv...
research
03/01/2023

Learning curves for deep structured Gaussian feature models

In recent years, significant attention in deep learning theory has been ...

Please sign up or login with your details

Forgot password? Click here to reset