Dynamic Regularizer with an Informative Prior

10/31/2019
by   Avinash Kori, et al.
0

Regularization methods, specifically those which directly alter weights like L_1 and L_2, are an integral part of many learning algorithms. Both the regularizers mentioned above are formulated by assuming certain priors in the parameter space and these assumptions, in some cases, induce sparsity in the parameter space. Regularizers help in transferring beliefs one has on the dataset or the parameter space by introducing adequate terms in the loss function. Any kind of formulation represents a specific set of beliefs: L_1 regularization conveys that the parameter space should be sparse whereas L_2 regularization conveys that the parameter space should be bounded and continuous. These regularizers in turn leverage certain priors to express these inherent beliefs. A better understanding of how the prior affects the behavior of the parameters and how the priors can be updated based on the dataset can contribute greatly in improving the generalization capabilities of a function estimator. In this work, we introduce a weakly informative prior and then further extend it to an informative prior in order to formulate a regularization penalty, which shows better results in terms of inducing sparsity experimentally, when compared to regularizers based only on Gaussian and Laplacian priors. Experimentally, we verify that a regularizer based on an adapted prior improves the generalization capabilities of any network. We illustrate the performance of the proposed method on the MNIST and CIFAR-10 datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/09/2014

Scoring and Searching over Bayesian Networks with Causal and Associative Priors

A significant theoretical advantage of search-and-score methods for lear...
research
09/18/2018

Comparison between Suitable Priors for Additive Bayesian Networks

Additive Bayesian networks are types of graphical models that extend the...
research
01/30/2019

Informative extended Mallows priors in the Bayesian Mallows model

The aim of this work is to study the problem of prior elicitation for th...
research
11/09/2020

Numerical Exploration of Training Loss Level-Sets in Deep Neural Networks

We present a computational method for empirically characterizing the tra...
research
12/07/2013

Robust Subspace System Identification via Weighted Nuclear Norm Optimization

Subspace identification is a classical and very well studied problem in ...
research
07/02/2019

Visual analytics for team-based invasion sports with significant events and Markov reward process

In team-based invasion sports such as soccer and basketball, analytics i...
research
07/28/2023

Optimal multi-environment causal regularization

In this manuscript we derive the optimal out-of-sample causal predictor ...

Please sign up or login with your details

Forgot password? Click here to reset