End-to-end Multiple Instance Learning with Gradient Accumulation

03/08/2022
by   Axel Andersson, et al.
0

Being able to learn on weakly labeled data, and provide interpretability, are two of the main reasons why attention-based deep multiple instance learning (ABMIL) methods have become particularly popular for classification of histopathological images. Such image data usually come in the form of gigapixel-sized whole-slide-images (WSI) that are cropped into smaller patches (instances). However, the sheer size of the data makes training of ABMIL models challenging. All the instances from one WSI cannot be processed at once by conventional GPUs. Existing solutions compromise training by relying on pre-trained models, strategic sampling or selection of instances, or self-supervised learning. We propose a training strategy based on gradient accumulation that enables direct end-to-end training of ABMIL models without being limited by GPU memory. We conduct experiments on both QMNIST and Imagenette to investigate the performance and training time, and compare with the conventional memory-expensive baseline and a recent sampled-based approach. This memory-efficient approach, although slower, reaches performance indistinguishable from the memory-expensive baseline.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/17/2021

Investigating a Baseline Of Self Supervised Learning Towards Reducing Labeling Costs For Image Classification

Data labeling in supervised learning is considered an expensive and infe...
research
08/04/2023

ES-MVSNet: Efficient Framework for End-to-end Self-supervised Multi-View Stereo

Compared to the multi-stage self-supervised multi-view stereo (MVS) meth...
research
03/22/2022

Remember Intentions: Retrospective-Memory-based Trajectory Prediction

To realize trajectory prediction, most previous methods adopt the parame...
research
11/09/2022

Efficient Speech Translation with Pre-trained Models

When building state-of-the-art speech translation models, the need for l...
research
11/19/2021

Fast and Data-Efficient Training of Rainbow: an Experimental Study on Atari

Across the Arcade Learning Environment, Rainbow achieves a level of perf...
research
10/24/2022

Iterative Patch Selection for High-Resolution Image Recognition

High-resolution images are prevalent in various applications, such as au...
research
05/17/2022

Moving Stuff Around: A study on efficiency of moving documents into memory for Neural IR models

When training neural rankers using Large Language Models, it's expected ...

Please sign up or login with your details

Forgot password? Click here to reset