'Rarely' a problem? Language models exhibit inverse scaling in their predictions following 'few'-type quantifiers

12/16/2022
by   James A. Michaelov, et al.
0

Language Models appear to perform poorly on quantification. We ask how badly. 'Few'-type quantifiers, as in 'few children like vegetables' might pose a particular challenge for Language Models, since the sentence components without the quantifier are likely to co-occur, and because 'few'-type quantifiers are rare. We present 960 sentences stimuli from two human neurolinguistic experiments to 22 autoregressive transformer models of differing sizes. Not only do the models perform poorly on 'few'-type quantifiers, but overall the larger the model, the worse its performance. We interpret this inverse scaling as suggesting that larger models increasingly reflect online rather than offline human processing, and argue that decreasing performance of larger models may challenge uses of Language Models as the basis for Natural Language Systems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/24/2018

Assessing Language Models with Scaling Properties

Language models have primarily been evaluated with perplexity. While per...
research
04/07/2022

Testing the limits of natural language models for predicting human language judgments

Neural network language models can serve as computational hypotheses abo...
research
06/12/2023

Probing Quantifier Comprehension in Large Language Models

With their increasing size, Large language models (LLMs) are becoming in...
research
08/30/2022

Do language models make human-like predictions about the coreferents of Italian anaphoric zero pronouns?

Some languages allow arguments to be omitted in certain contexts. Yet hu...
research
05/27/2023

Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models

Language models have been shown to exhibit positive scaling, where perfo...
research
06/15/2023

Inverse Scaling: When Bigger Isn't Better

Work on scaling laws has found that large language models (LMs) show pre...
research
05/24/2023

Emergent inabilities? Inverse scaling over the course of pretraining

Does inverse scaling only occur as a function of model parameter size, o...

Please sign up or login with your details

Forgot password? Click here to reset