Analyzing the Source and Target Contributions to Predictions in Neural Machine Translation

10/21/2020
by   Elena Voita, et al.
7

In Neural Machine Translation (and, more generally, conditional language modeling), the generation of a target token is influenced by two types of context: the source and the prefix of the target sequence. While many attempts to understand the internal workings of NMT models have been made, none of them explicitly evaluates relative source and target contributions to a generation decision. We argue that this relative contribution can be evaluated by adopting a variant of Layerwise Relevance Propagation (LRP). Its underlying 'conservation principle' makes relevance propagation unique: differently from other methods, it evaluates not an abstract quantity reflecting token importance, but the proportion of each token's influence. We extend LRP to the Transformer and conduct an analysis of NMT models which explicitly evaluates the source and target relative contributions to the generation process. We analyze changes in these contributions when conditioning on different types of prefixes, when varying the training objective or the amount of training data, and during the training process. We find that models trained with more data tend to rely on source information more and to have more sharp token contributions; the training process is non-monotonic with several stages of different nature.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2022

Towards Opening the Black Box of Neural Machine Translation: Source and Target Interpretations of the Transformer

In Neural Machine Translation (NMT), each token prediction is conditione...
research
08/22/2016

Context Gates for Neural Machine Translation

In neural machine translation (NMT), generation of a target word depends...
research
09/03/2021

Language Modeling, Lexical Translation, Reordering: The Training Process of NMT through the Lens of Classical SMT

Differently from the traditional statistical MT that decomposes the tran...
research
05/31/2021

Beyond Noise: Mitigating the Impact of Fine-grained Semantic Divergences on Neural Machine Translation

While it has been shown that Neural Machine Translation (NMT) is highly ...
research
08/29/2019

Regularized Context Gates on Transformer for Machine Translation

Context gates are effective to control the contributions from the source...
research
01/18/2023

Understanding and Detecting Hallucinations in Neural Machine Translation via Model Introspection

Neural sequence generation models are known to "hallucinate", by produci...
research
04/19/2022

Generating Authentic Adversarial Examples beyond Meaning-preserving with Doubly Round-trip Translation

Generating adversarial examples for Neural Machine Translation (NMT) wit...

Please sign up or login with your details

Forgot password? Click here to reset