Quiver neural networks

07/26/2022
by   Iordan Ganev, et al.
0

We develop a uniform theoretical approach towards the analysis of various neural network connectivity architectures by introducing the notion of a quiver neural network. Inspired by quiver representation theory in mathematics, this approach gives a compact way to capture elaborate data flows in complex network architectures. As an application, we use parameter space symmetries to prove a lossless model compression algorithm for quiver neural networks with certain non-pointwise activations known as rescaling activations. In the case of radial rescaling activations, we prove that training the compressed model with gradient descent is equivalent to training the original model with projected gradient descent.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/06/2021

The QR decomposition for radial neural networks

We provide a theoretical framework for neural networks in terms of the r...
research
08/25/2019

What are Neural Networks made of?

The success of Deep Learning methods is not well understood, though vari...
research
12/04/2021

Optimization-Based Separations for Neural Networks

Depth separation results propose a possible theoretical explanation for ...
research
02/12/2019

Towards moderate overparameterization: global convergence guarantees for training shallow neural networks

Many modern neural network architectures are trained in an overparameter...
research
06/22/2020

Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent

We prove the first superpolynomial lower bounds for learning one-layer n...
research
12/14/2017

Nonparametric Neural Networks

Automatically determining the optimal size of a neural network for a giv...
research
11/14/2020

GENNI: Visualising the Geometry of Equivalences for Neural Network Identifiability

We propose an efficient algorithm to visualise symmetries in neural netw...

Please sign up or login with your details

Forgot password? Click here to reset