Conceptor-Aided Debiasing of Contextualized Embeddings

11/20/2022
by   Yifei Li, et al.
0

Pre-trained language models reflect the inherent social biases of their training corpus. Many methods have been proposed to mitigate this issue, but they often fail to debias or they sacrifice model accuracy. We use conceptors–a soft projection method–to identify and remove the bias subspace in contextual embeddings in BERT and GPT. We propose two methods of applying conceptors (1) bias subspace projection by post-processing; and (2) a new architecture, conceptor-intervened BERT (CI-BERT), which explicitly incorporates the conceptor projection into all layers during training. We find that conceptor post-processing achieves state-of-the-art debiasing results while maintaining or improving BERT's performance on the GLUE benchmark. Although CI-BERT's training takes all layers' bias into account and can outperform its post-processing counterpart in bias mitigation, CI-BERT reduces the language model accuracy. We also show the importance of carefully constructing the bias subspace. The best results are obtained by removing outliers from the list of biased words, intersecting them (using the conceptor AND operation), and computing their embeddings using the sentences from a cleaner corpus.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2020

Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation

Word embeddings derived from human-generated corpora inherit strong gend...
research
04/15/2021

Effect of Post-processing on Contextualized Word Representations

Post-processing of static embedding has beenshown to improve their perfo...
research
09/30/2020

Interactive Re-Fitting as a Technique for Improving Word Embeddings

Word embeddings are a fixed, distributional representation of the contex...
research
09/10/2020

Investigating Gender Bias in BERT

Contextual language models (CLMs) have pushed the NLP benchmarks to a ne...
research
02/16/2020

Towards Detection of Subjective Bias using Contextualized Word Embeddings

Subjective bias detection is critical for applications like propaganda d...
research
09/16/2022

Less is Better: Recovering Intended-Feature Subspace to Robustify NLU Models

Datasets with significant proportions of bias present threats for traini...
research
11/10/2022

ADEPT: A DEbiasing PrompT Framework

Several works have proven that finetuning is an applicable approach for ...

Please sign up or login with your details

Forgot password? Click here to reset