LMPriors: Pre-Trained Language Models as Task-Specific Priors

10/22/2022
by   Kristy Choi, et al.
9

Particularly in low-data regimes, an outstanding challenge in machine learning is developing principled techniques for augmenting our models with suitable priors. This is to encourage them to learn in ways that are compatible with our understanding of the world. But in contrast to generic priors such as shrinkage or sparsity, we draw inspiration from the recent successes of large-scale language models (LMs) to construct task-specific priors distilled from the rich knowledge of LMs. Our method, Language Model Priors (LMPriors), incorporates auxiliary natural language metadata about the task – such as variable names and descriptions – to encourage downstream model outputs to be consistent with the LM's common-sense reasoning based on the metadata. Empirically, we demonstrate that LMPriors improve model performance in settings where such natural language descriptions are available, and perform well on several tasks that benefit from such prior knowledge, such as feature selection, causal inference, and safe reinforcement learning.

READ FULL TEXT
research
09/20/2021

Learning Natural Language Generation from Scratch

This paper introduces TRUncated ReinForcement Learning for Language (Tru...
research
12/16/2021

CLIN-X: pre-trained language models and a study on cross-task transfer for concept extraction in the clinical domain

The field of natural language processing (NLP) has recently seen a large...
research
08/26/2022

Task-specific Pre-training and Prompt Decomposition for Knowledge Graph Population with Language Models

We present a system for knowledge graph population with Language Models,...
research
07/24/2023

Making Metadata More FAIR Using Large Language Models

With the global increase in experimental data artifacts, harnessing them...
research
07/21/2023

Selective Perception: Optimizing State Descriptions with Reinforcement Learning for Language Model Actors

Large language models (LLMs) are being applied as actors for sequential ...
research
04/25/2023

What's in a Name? Evaluating Assembly-Part Semantic Knowledge in Language Models through User-Provided Names in CAD Files

Semantic knowledge of part-part and part-whole relationships in assembli...
research
03/29/2023

Personalised Language Modelling of Screen Characters Using Rich Metadata Annotations

Personalisation of language models for dialogue sensitises them to bette...

Please sign up or login with your details

Forgot password? Click here to reset