Learning De-biased Representations with Biased Representations

10/07/2019
by   Hyojin Bahng, et al.
25

Many machine learning algorithms are trained and evaluated by splitting data from a single source into training and test sets. While such focus on in-distribution learning scenarios has led interesting advances, it has not been able to tell if models are relying on dataset biases as shortcuts for successful prediction (e.g., using snow cues for recognising snowmobiles). Such biased models fail to generalise when the bias shifts to a different class. The cross-bias generalisation problem has been addressed by de-biasing training data through augmentation or re-sampling, which are often prohibitive due to the data collection cost (e.g., collecting images of a snowmobile on a desert) and the difficulty of quantifying or expressing biases in the first place. In this work, we propose a novel framework to train a de-biased representation by encouraging it to be different from a set of representations that are biased by design. This tactic is feasible in many scenarios where it is much easier to define a set of biased representations than to define and quantify bias. Our experiments and analyses show that our method discourages models from taking bias shortcuts, resulting in improved performances on de-biased test data.

READ FULL TEXT

page 1

page 3

page 4

page 5

page 6

page 8

page 10

page 11

research
05/30/2023

Fighting Bias with Bias: Promoting Model Robustness by Amplifying Dataset Biases

NLP models often rely on superficial cues known as dataset biases to ach...
research
08/23/2021

Exploring Biases and Prejudice of Facial Synthesis via Semantic Latent Space

Deep learning (DL) models are widely used to provide a more convenient a...
research
11/19/2022

Quantifying Human Bias and Knowledge to guide ML models during Training

This paper discusses a crowdsourcing based method that we designed to qu...
research
03/26/2021

Mixing-AdaSIN: Constructing a de-biased dataset using Adaptive Structural Instance Normalization and texture Mixing

Following the pandemic outbreak, several works have proposed to diagnose...
research
10/15/2021

Data Generation using Texture Co-occurrence and Spatial Self-Similarity for Debiasing

Classification models trained on biased datasets usually perform poorly ...
research
09/28/2020

Why resampling outperforms reweighting for correcting sampling bias

A data set sampled from a certain population is biased if the subgroups ...
research
04/28/2022

Learning to Split for Automatic Bias Detection

Classifiers are biased when trained on biased datasets. As a remedy, we ...

Please sign up or login with your details

Forgot password? Click here to reset