On the Inherent Regularization Effects of Noise Injection During Training

02/15/2021
by   Oussama Dhifallah, et al.
0

Randomly perturbing networks during the training process is a commonly used approach to improving generalization performance. In this paper, we present a theoretical study of one particular way of random perturbation, which corresponds to injecting artificial noise to the training data. We provide a precise asymptotic characterization of the training and generalization errors of such randomly perturbed learning problems on a random feature model. Our analysis shows that Gaussian noise injection in the training process is equivalent to introducing a weighted ridge regularization, when the number of noise injections tends to infinity. The explicit form of the regularization is also given. Numerical results corroborate our asymptotic predictions, showing that they are accurate even in moderate problem dimensions. Our theoretical predictions are based on a new correlated Gaussian equivalence conjecture that generalizes recent results in the study of random feature models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/27/2020

A Precise Performance Analysis of Learning with Random Features

We study the problem of learning an unknown function using random featur...
research
09/27/2018

Introducing Noise in Decentralized Training of Neural Networks

It has been shown that injecting noise into the neural network weights d...
research
06/05/2019

Neural SDE: Stabilizing Neural ODE Networks with Stochastic Noise

Neural Ordinary Differential Equation (Neural ODE) has been proposed as ...
research
06/08/2022

On gradient descent training under data augmentation with on-line noisy copies

In machine learning, data augmentation (DA) is a technique for improving...
research
09/16/2020

Universality Laws for High-Dimensional Learning with Random Features

We prove a universality theorem for learning with random features. Our r...
research
07/25/2019

Adaptive Noise Injection: A Structure-Expanding Regularization for RNN

The vanilla LSTM has become one of the most potential architectures in w...
research
12/05/2020

Optimal polymer slugs injection curves

In the paper we study miscible injection of liquids with decreasing visc...

Please sign up or login with your details

Forgot password? Click here to reset