Do We Need Fully Connected Output Layers in Convolutional Networks?

04/28/2020
by   Zhongchao Qian, et al.
46

Traditionally, deep convolutional neural networks consist of a series of convolutional and pooling layers followed by one or more fully connected (FC) layers to perform the final classification. While this design has been successful, for datasets with a large number of categories, the fully connected layers often account for a large percentage of the network's parameters. For applications with memory constraints, such as mobile devices and embedded platforms, this is not ideal. Recently, a family of architectures that involve replacing the learned fully connected output layer with a fixed layer has been proposed as a way to achieve better efficiency. In this paper we examine this idea further and demonstrate that fixed classifiers offer no additional benefit compared to simply removing the output layer along with its parameters. We further demonstrate that the typical approach of having a fully connected final output layer is inefficient in terms of parameter count. We are able to achieve comparable performance to a traditionally learned fully connected classification output layer on the ImageNet-1K, CIFAR-100, Stanford Cars-196, and Oxford Flowers-102 datasets, while not having a fully connected output layer at all.

READ FULL TEXT

page 1

page 3

research
12/22/2014

Deep Fried Convnets

The fully connected layers of a deep convolutional neural network typica...
research
11/19/2018

An Efficient Transfer Learning Technique by Using Final Fully-Connected Layer Output Features of Deep Networks

In this paper, we propose a computationally efficient transfer learning ...
research
10/22/2021

Deep Convolutional Autoencoders as Generic Feature Extractors in Seismological Applications

The idea of using a deep autoencoder to encode seismic waveform features...
research
06/05/2019

Visual Confusion Label Tree For Image Classification

Convolution neural network models are widely used in image classificatio...
research
12/27/2021

Learning Robust and Lightweight Model through Separable Structured Transformations

With the proliferation of mobile devices and the Internet of Things, dee...
research
07/23/2020

WeightNet: Revisiting the Design Space of Weight Networks

We present a conceptually simple, flexible and effective framework for w...
research
02/17/2021

Beyond Fully-Connected Layers with Quaternions: Parameterization of Hypercomplex Multiplications with 1/n Parameters

Recent works have demonstrated reasonable success of representation lear...

Please sign up or login with your details

Forgot password? Click here to reset