Training Binary Neural Networks using the Bayesian Learning Rule

02/25/2020
by   Xiangming Meng, et al.
14

Neural networks with binary weights are computation-efficient and hardware-friendly, but their training is challenging because it involves a discrete optimization problem. Surprisingly, ignoring the discrete nature of the problem and using gradient-based methods, such as Straight-Through Estimator, still works well in practice. This raises the question: are there principled approaches which justify such methods? In this paper, we propose such an approach using the Bayesian learning rule. The rule, when applied to estimate a Bernoulli distribution over the binary weights, results in an algorithm which justifies some of the algorithmic choices made by the previous approaches. The algorithm not only obtains state-of-the-art performance, but also enables uncertainty estimation for continual learning to avoid catastrophic forgetting. Our work provides a principled approach for training binary neural networks which justifies and extends existing approaches.

READ FULL TEXT
research
06/06/2019

Uncertainty-guided Continual Learning with Bayesian Neural Networks

Continual learning aims to learn new tasks without forgetting previously...
research
01/04/2023

On Sequential Bayesian Inference for Continual Learning

Sequential Bayesian inference can be used for continual learning to prev...
research
12/06/2021

AdaSTE: An Adaptive Straight-Through Estimator to Train Binary Neural Networks

We propose a new algorithm for training deep neural networks (DNNs) with...
research
03/27/2022

Continual learning: a feature extraction formalization, an efficient algorithm, and fundamental obstructions

Continual learning is an emerging paradigm in machine learning, wherein ...
research
06/11/2020

Reintroducing Straight-Through Estimators as Principled Methods for Stochastic Binary Networks

Training neural networks with binary weights and activations is a challe...
research
10/21/2017

Learning Discrete Weights Using the Local Reparameterization Trick

Recent breakthroughs in computer vision make use of large deep neural ne...
research
03/27/2013

The Automatic Training of Rule Bases that Use Numerical Uncertainty Representations

The use of numerical uncertainty representations allows better modeling ...

Please sign up or login with your details

Forgot password? Click here to reset