Retrospective Loss: Looking Back to Improve Training of Deep Neural Networks

06/24/2020
by   Surgan Jandial, et al.
0

Deep neural networks (DNNs) are powerful learning machines that have enabled breakthroughs in several domains. In this work, we introduce a new retrospective loss to improve the training of deep neural network models by utilizing the prior experience available in past model states during training. Minimizing the retrospective loss, along with the task-specific loss, pushes the parameter state at the current training step towards the optimal parameter state while pulling it away from the parameter state at a previous training step. Although a simple idea, we analyze the method as well as to conduct comprehensive sets of experiments across domains - images, speech, text, and graphs - to show that the proposed loss results in improved performance across input domains, tasks, and architectures.

READ FULL TEXT
research
04/26/2020

COLAM: Co-Learning of Deep Neural Networks and Soft Labels via Alternating Minimization

Softening labels of training datasets with respect to data representatio...
research
10/07/2018

Principled Deep Neural Network Training through Linear Programming

Deep Learning has received significant attention due to its impressive p...
research
04/12/2019

An Empirical Evaluation Study on the Training of SDC Features for Dense Pixel Matching

Training a deep neural network is a non-trivial task. Not only the tunin...
research
12/27/2022

MixupE: Understanding and Improving Mixup from Directional Derivative Perspective

Mixup is a popular data augmentation technique for training deep neural ...
research
01/31/2018

Optimizing Non-decomposable Measures with Deep Networks

We present a class of algorithms capable of directly training deep neura...
research
05/18/2018

Reconstruction of training samples from loss functions

This paper presents a new mathematical framework to analyze the loss fun...
research
10/06/2021

Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space Perspective

Deep neural networks (DNNs) often rely on easy-to-learn discriminatory f...

Please sign up or login with your details

Forgot password? Click here to reset