Look-ahead Attention for Generation in Neural Machine Translation

08/30/2017
by   Long Zhou, et al.
0

The attention model has become a standard component in neural machine translation (NMT) and it guides translation process by selectively focusing on parts of the source sentence when predicting each target word. However, we find that the generation of a target word does not only depend on the source sentence, but also rely heavily on the previous generated target words, especially the distant words which are difficult to model by using recurrent neural networks. To solve this problem, we propose in this paper a novel look-ahead attention mechanism for generation in NMT, which aims at directly capturing the dependency relationship between target words. We further design three patterns to integrate our look-ahead attention into the conventional attention model. Experiments on NIST Chinese-to-English and WMT English-to-German translation tasks show that our proposed look-ahead attention mechanism achieves substantial improvements over state-of-the-art baselines.

READ FULL TEXT
research
07/04/2017

An empirical study on the effectiveness of images in Multimodal Neural Machine Translation

In state-of-the-art Neural Machine Translation (NMT), an attention mecha...
research
05/31/2017

Learning When to Attend for Neural Machine Translation

In the past few years, attention mechanisms have become an indispensable...
research
03/08/2020

Investigating the Decoders of Maximum Likelihood Sequence Models: A Look-ahead Approach

We demonstrate how we can practically incorporate multi-step future info...
research
11/12/2017

Syntax-Directed Attention for Neural Machine Translation

Attention mechanism, including global attention and local attention, pla...
research
07/18/2016

Neural Machine Translation with Recurrent Attention Modeling

Knowing which words have been attended to in previous time steps while g...
research
09/30/2019

Interrogating the Explanatory Power of Attention in Neural Machine Translation

Attention models have become a crucial component in neural machine trans...
research
02/04/2020

Syntactically Look-Ahead Attention Network for Sentence Compression

Sentence compression is the task of compressing a long sentence into a s...

Please sign up or login with your details

Forgot password? Click here to reset