Residual networks classify inputs based on their neural transient dynamics

01/08/2021
by   Fereshteh Lagzi, et al.
42

In this study, we analyze the input-output behavior of residual networks from a dynamical system point of view by disentangling the residual dynamics from the output activities before the classification stage. For a network with simple skip connections between every successive layer, and for logistic activation function, and shared weights between layers, we show analytically that there is a cooperation and competition dynamics between residuals corresponding to each input dimension. Interpreting these kind of networks as nonlinear filters, the steady state value of the residuals in the case of attractor networks are indicative of the common features between different input dimensions that the network has observed during training, and has encoded in those components. In cases where residuals do not converge to an attractor state, their internal dynamics are separable for each input class, and the network can reliably approximate the output. We bring analytical and empirical evidence that residual networks classify inputs based on the integration of the transient dynamics of the residuals, and will show how the network responds to input perturbations. We compare the network dynamics for a ResNet and a Multi-Layer Perceptron and show that the internal dynamics, and the noise evolution are fundamentally different in these networks, and ResNets are more robust to noisy inputs. Based on these findings, we also develop a new method to adjust the depth for residual networks during training. As it turns out, after pruning the depth of a ResNet using this algorithm,the network is still capable of classifying inputs with a high accuracy.

READ FULL TEXT

page 8

page 16

research
01/07/2020

Kinetic Theory for Residual Neural Networks

Deep residual neural networks (ResNet) are performing very well for many...
research
08/04/2021

A Method to use Nonlinear Dynamics in a Whisker Sensor for Terrain Identification by Mobile Robots

This paper shows analytical and experimental evidence of using the vibra...
research
02/19/2021

Training cascaded networks for speeded decisions using a temporal-difference loss

Although deep feedforward neural networks share some characteristics wit...
research
05/31/2019

Residual Networks as Nonlinear Systems: Stability Analysis using Linearization

We regard pre-trained residual networks (ResNets) as nonlinear systems a...
research
09/15/2022

BadRes: Reveal the Backdoors through Residual Connection

Generally, residual connections are indispensable network components in ...
research
12/24/2017

Mean Field Residual Networks: On the Edge of Chaos

We study randomly initialized residual networks using mean field theory ...

Please sign up or login with your details

Forgot password? Click here to reset