Linguistic Knowledge as Memory for Recurrent Neural Networks

03/07/2017
by   Bhuwan Dhingra, et al.
0

Training recurrent neural networks to model long term dependencies is difficult. Hence, we propose to use external linguistic knowledge as an explicit signal to inform the model which memories it should utilize. Specifically, external knowledge is used to augment a sequence with typed edges between arbitrarily distant elements, and the resulting graph is decomposed into directed acyclic subgraphs. We introduce a model that encodes such graphs as explicit memory in recurrent neural networks, and use it to model coreference relations in text. We apply our model to several text comprehension tasks and achieve new state-of-the-art results on all considered benchmarks, including CNN, bAbi, and LAMBADA. On the bAbi QA tasks, our model solves 15 out of the 20 tasks with only 1000 training examples per task. Analysis of the learned representations further demonstrates the ability of our model to encode fine-grained entity information across a document.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2015

Recurrent Neural Networks with External Memory for Language Understanding

Recurrent Neural Networks (RNNs) have become increasingly popular for th...
research
06/12/2018

Quaternion Recurrent Neural Networks

Recurrent neural networks (RNNs) are powerful architectures to model seq...
research
06/02/2021

Warming-up recurrent neural networks to maximize reachable multi-stability greatly improves learning

Training recurrent neural networks is known to be difficult when time de...
research
02/25/2019

Leveraging Knowledge Bases in LSTMs for Improving Machine Reading

This paper focuses on how to take advantage of external knowledge bases ...
research
04/09/2016

Towards an Indexical Model of Situated Language Comprehension for Cognitive Agents in Physical Worlds

We propose a computational model of situated language comprehension base...
research
02/15/2017

Generative Temporal Models with Memory

We consider the general problem of modeling temporal data with long-rang...
research
04/16/2018

Neural Models for Reasoning over Multiple Mentions using Coreference

Many problems in NLP require aggregating information from multiple menti...

Please sign up or login with your details

Forgot password? Click here to reset