Continual learning using hash-routed convolutional neural networks

10/09/2020
by   Ahmad Berjaoui, et al.
0

Continual learning could shift the machine learning paradigm from data centric to model centric. A continual learning model needs to scale efficiently to handle semantically different datasets, while avoiding unnecessary growth. We introduce hash-routed convolutional neural networks: a group of convolutional units where data flows dynamically. Feature maps are compared using feature hashing and similar data is routed to the same units. A hash-routed network provides excellent plasticity thanks to its routed nature, while generating stable features through the use of orthogonal feature hashing. Each unit evolves separately and new units can be added (to be used only when necessary). Hash-routed networks achieve excellent performance across a variety of typical continual learning benchmarks without storing raw data and train using only gradient descent. Besides providing a continual learning framework for supervised tasks with encouraging results, our model can be used for unsupervised or reinforcement learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2020

Adaptive Group Sparse Regularization for Continual Learning

We propose a novel regularization-based continual learning method, dubbe...
research
08/13/2021

Continual Backprop: Stochastic Gradient Descent with Persistent Randomness

The Backprop algorithm for learning in neural networks utilizes two mech...
research
06/04/2023

Towards Robust Feature Learning with t-vFM Similarity for Continual Learning

Continual learning has been developed using standard supervised contrast...
research
12/08/2019

Nonparametric Bayesian Structure Adaptation for Continual Learning

Continual Learning is a learning paradigm where machine learning models ...
research
06/17/2022

Debugging using Orthogonal Gradient Descent

In this report we consider the following problem: Given a trained model ...
research
05/28/2018

Parallel Weight Consolidation: A Brain Segmentation Case Study

Collecting the large datasets needed to train deep neural networks can b...
research
03/20/2019

Convolution with even-sized kernels and symmetric padding

Compact convolutional neural networks gain efficiency mainly through dep...

Please sign up or login with your details

Forgot password? Click here to reset