Maxout Networks

02/18/2013
by   Ian J. Goodfellow, et al.
0

We consider the problem of designing models to leverage a recently introduced approximate model averaging technique called dropout. We define a simple new model called maxout (so named because its output is the max of a set of inputs, and because it is a natural companion to dropout) designed to both facilitate optimization by dropout and improve the accuracy of dropout's fast approximate model averaging technique. We empirically verify that the model successfully accomplishes both of these tasks. We use maxout and dropout to demonstrate state of the art classification performance on four benchmark datasets: MNIST, CIFAR-10, CIFAR-100, and SVHN.

READ FULL TEXT
research
12/01/2015

Towards Dropout Training for Convolutional Neural Networks

Recently, dropout has seen increasing use in deep learning. For deep con...
research
04/11/2021

UniDrop: A Simple yet Effective Technique to Improve Transformer without Extra Cost

Transformer architecture achieves great success in abundant natural lang...
research
06/23/2019

Confidence Calibration for Convolutional Neural Networks Using Structured Dropout

In classification applications, we often want probabilistic predictions ...
research
12/20/2013

Improving Deep Neural Networks with Probabilistic Maxout Units

We present a probabilistic variant of the recently introduced maxout uni...
research
05/23/2018

Pushing the bounds of dropout

We show that dropout training is best understood as performing MAP estim...
research
11/19/2019

Thick-Net: Parallel Network Structure for Sequential Modeling

Recurrent neural networks have been widely used in sequence learning tas...
research
02/26/2022

Dropout can Simulate Exponential Number of Models for Sample Selection Techniques

Following Coteaching, generally in the literature, two models are used i...

Please sign up or login with your details

Forgot password? Click here to reset