UDALM: Unsupervised Domain Adaptation through Language Modeling

04/14/2021
by   Constantinos Karouzos, et al.
0

In this work we explore Unsupervised Domain Adaptation (UDA) of pretrained language models for downstream tasks. We introduce UDALM, a fine-tuning procedure, using a mixed classification and Masked Language Model loss, that can adapt to the target domain distribution in a robust and sample efficient manner. Our experiments show that performance of models trained with the mixed loss scales with the amount of available target data and the mixed loss can be effectively used as a stopping criterion during UDA training. Furthermore, we discuss the relationship between A-distance and the target error and explore some limitations of the Domain Adversarial Training approach. Our method is evaluated on twelve domain pairs of the Amazon Reviews Sentiment dataset, yielding 91.74% accuracy, which is an 1.11% absolute improvement over the state-of-the-art.

READ FULL TEXT

page 3

page 7

research
05/10/2019

Virtual Mixup Training for Unsupervised Domain Adaptation

We study the problem of unsupervised domain adaptation which aims to ada...
research
09/24/2020

Feature Adaptation of Pre-Trained Language Models across Languages and Domains for Text Classification

Adapting pre-trained language models (PrLMs) (e.g., BERT) to new domains...
research
09/04/2019

Distributionally Robust Language Modeling

Language models are generally trained on data spanning a wide range of t...
research
10/05/2020

Effective Unsupervised Domain Adaptation with Adversarially Trained Language Models

Recent work has shown the importance of adaptation of broad-coverage con...
research
11/06/2022

On the Domain Adaptation and Generalization of Pretrained Language Models: A Survey

Recent advances in NLP are brought by a range of large-scale pretrained ...
research
12/16/2021

Efficient Hierarchical Domain Adaptation for Pretrained Language Models

Generative language models are trained on diverse, general domain corpor...
research
05/03/2022

Mixed-effects transformers for hierarchical adaptation

Language use differs dramatically from context to context. To some degre...

Please sign up or login with your details

Forgot password? Click here to reset