Dynamic Neural Diversification: Path to Computationally Sustainable Neural Networks

09/20/2021
by   Alexander Kovalenko, et al.
6

Small neural networks with a constrained number of trainable parameters, can be suitable resource-efficient candidates for many simple tasks, where now excessively large models are used. However, such models face several problems during the learning process, mainly due to the redundancy of the individual neurons, which results in sub-optimal accuracy or the need for additional training steps. Here, we explore the diversity of the neurons within the hidden layer during the learning process, and analyze how the diversity of the neurons affects predictions of the model. As following, we introduce several techniques to dynamically reinforce diversity between neurons during the training. These decorrelation techniques improve learning at early stages and occasionally help to overcome local minima faster. Additionally, we describe novel weight initialization method to obtain decorrelated, yet stochastic weight initialization for a fast and efficient neural network training. Decorrelated weight initialization in our case shows about 40 accuracy during the first 5 epochs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/02/2020

Persistent Neurons

Most algorithms used in neural networks(NN)-based leaning tasks are stro...
research
12/04/2018

Parameter Re-Initialization through Cyclical Batch Size Schedules

Optimal parameter initialization remains a crucial problem for neural ne...
research
02/17/2022

When, where, and how to add new neurons to ANNs

Neurogenesis in ANNs is an understudied and difficult problem, even comp...
research
02/28/2022

How and what to learn:The modes of machine learning

We proposal a new approach, namely the weight pathway analysis (WPA), to...
research
11/16/2022

Engineering Monosemanticity in Toy Models

In some neural networks, individual neurons correspond to natural “featu...
research
11/25/2022

Bypass Exponential Time Preprocessing: Fast Neural Network Training via Weight-Data Correlation Preprocessing

Over the last decade, deep neural networks have transformed our society,...
research
08/02/2019

Network with Sub-Networks

We introduce network with sub-network, a neural network which their weig...

Please sign up or login with your details

Forgot password? Click here to reset