Join-Chain Network: A Logical Reasoning View of the Multi-head Attention in Transformer

10/06/2022
by   Jianyi Zhang, et al.
0

Developing neural architectures that are capable of logical reasoning has become increasingly important for a wide range of applications (e.g., natural language processing). Towards this grand objective, we propose a symbolic reasoning architecture that chains many join operators together to model output logical expressions. In particular, we demonstrate that such an ensemble of join-chains can express a broad subset of ”tree-structured” first-order logical expressions, named FOET, which is particularly useful for modeling natural languages. To endow it with differentiable learning capability, we closely examine various neural operators for approximating the symbolic join-chains. Interestingly, we find that the widely used multi-head self-attention module in transformer can be understood as a special neural operator that implements the union bound of the join operator in probabilistic predicate space. Our analysis not only provides a new perspective on the mechanism of the pretrained models such as BERT for natural language understanding but also suggests several important future improvement directions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2023

Learning Language Representations with Logical Inductive Bias

Transformer architectures have achieved great success in solving natural...
research
10/01/2021

Natural language understanding for logical games

We developed a system able to automatically solve logical puzzles in nat...
research
03/21/2023

Logical Reasoning over Natural Language as Knowledge Representation: A Survey

Logical reasoning is central to human cognition and intelligence. Past r...
research
09/20/2020

Repulsive Attention: Rethinking Multi-head Attention as Bayesian Inference

The neural attention mechanism plays an important role in many natural l...
research
06/21/2021

A Logical Model for joining Property Graphs

The present paper upgrades the logical model required to exploit materia...
research
05/10/2019

Using syntactical and logical forms to evaluate textual inference competence

In the light of recent breakthroughs in transfer learning for Natural La...
research
06/06/2019

Understanding and Improving Transformer From a Multi-Particle Dynamic System Point of View

The Transformer architecture is widely used in natural language processi...

Please sign up or login with your details

Forgot password? Click here to reset