Studying word order through iterative shuffling

09/10/2021
by   Nikolay Malkin, et al.
6

As neural language models approach human performance on NLP benchmark tasks, their advances are widely seen as evidence of an increasingly complex understanding of syntax. This view rests upon a hypothesis that has not yet been empirically tested: that word order encodes meaning essential to performing these tasks. We refute this hypothesis in many cases: in the GLUE suite and in various genres of English text, the words in a sentence or phrase can rarely be permuted to form a phrase carrying substantially different information. Our surprising result relies on inference by iterative shuffling (IBIS), a novel, efficient procedure that finds the ordering of a bag of words having the highest likelihood under a fixed language model. IBIS can use any black-box model without additional training and is superior to existing word ordering algorithms. Coalescing our findings, we discuss how shuffling inference procedures such as IBIS can benefit language modeling and constrained generation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/15/2022

On the Role of Pre-trained Language Models in Word Ordering: A Case Study with BART

Word ordering is a constrained language generation task taking unordered...
research
09/10/2021

Euphemistic Phrase Detection by Masked Language Model

It is a well-known approach for fringe groups and organizations to use e...
research
07/29/2021

Demystifying Neural Language Models' Insensitivity to Word-Order

Recent research analyzing the sensitivity of natural language understand...
research
09/18/2018

Analysis of Bag-of-n-grams Representation's Properties Based on Textual Reconstruction

Despite its simplicity, bag-of-n-grams sen- tence representation has bee...
research
08/05/2017

A Comparison of Neural Models for Word Ordering

We compare several language models for the word-ordering task and propos...
research
04/15/2021

Syntactic Perturbations Reveal Representational Correlates of Hierarchical Phrase Structure in Pretrained Language Models

While vector-based language representations from pretrained language mod...
research
10/17/2021

Schrödinger's Tree – On Syntax and Neural Language Models

In the last half-decade, the field of natural language processing (NLP) ...

Please sign up or login with your details

Forgot password? Click here to reset