Position Information in Transformers: An Overview

02/22/2021
by   Philipp Dufter, et al.
0

Transformers are arguably the main workhorse in recent Natural Language Processing research. By definition a Transformer is invariant with respect to reorderings of the input. However, language is inherently sequential and word order is essential to the semantics and syntax of an utterance. In this paper, we provide an overview of common methods to incorporate position information into Transformer models. The objectives of this survey are to i) showcase that position information in Transformer is a vibrant and extensive research area; ii) enable the reader to compare existing methods by providing a unified notation and meaningful clustering; iii) indicate what characteristics of an application should be taken into account when selecting a position encoding; iv) provide stimuli for future research.

READ FULL TEXT

page 8

page 15

research
09/14/2020

Efficient Transformers: A Survey

Transformer model architectures have garnered immense interest lately du...
research
02/28/2023

A Survey on Long Text Modeling with Transformers

Modeling long texts has been an essential technique in the field of natu...
research
11/08/2022

Word Order Matters when you Increase Masking

Word order, an essential property of natural languages, is injected in T...
research
06/08/2021

A Survey of Transformers

Transformers have achieved great success in many artificial intelligence...
research
07/06/2023

Vision Language Transformers: A Survey

Vision language tasks, such as answering questions about or generating c...
research
02/02/2023

A Survey on Efficient Training of Transformers

Recent advances in Transformers have come with a huge requirement on com...
research
05/02/2021

Synthesizing Abstract Transformers

This paper addresses the problem of creating abstract transformers autom...

Please sign up or login with your details

Forgot password? Click here to reset