Glancing Future for Simultaneous Machine Translation

09/12/2023
by   Shoutao Guo, et al.
0

Simultaneous machine translation (SiMT) outputs translation while reading the source sentence. Unlike conventional sequence-to-sequence (seq2seq) training, existing SiMT methods adopt the prefix-to-prefix (prefix2prefix) training, where the model predicts target tokens based on partial source tokens. However, the prefix2prefix training diminishes the ability of the model to capture global information and introduces forced predictions due to the absence of essential source information. Consequently, it is crucial to bridge the gap between the prefix2prefix training and seq2seq training to enhance the translation capability of the SiMT model. In this paper, we propose a novel method that glances future in curriculum learning to achieve the transition from the seq2seq training to prefix2prefix training. Specifically, we gradually reduce the available source information from the whole sentence to the prefix corresponding to that latency. Our method is applicable to a wide range of SiMT methods and experiments demonstrate that our method outperforms strong baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/20/2022

Wait-info Policy: Balancing Source and Target at Information Level for Simultaneous Machine Translation

Simultaneous machine translation (SiMT) outputs the translation while re...
research
05/18/2020

Efficient Wait-k Models for Simultaneous Machine Translation

Simultaneous machine translation consists in starting output generation ...
research
04/27/2022

Data-Driven Adaptive Simultaneous Machine Translation

In simultaneous translation (SimulMT), the most widely used strategy is ...
research
03/17/2022

Reducing Position Bias in Simultaneous Machine Translation with Length-Aware Framework

Simultaneous machine translation (SiMT) starts translating while receivi...
research
03/01/2023

Hidden Markov Transformer for Simultaneous Machine Translation

Simultaneous machine translation (SiMT) outputs the target sequence whil...
research
12/12/2018

Sentence-wise Smooth Regularization for Sequence to Sequence Learning

Maximum-likelihood estimation (MLE) is widely used in sequence to sequen...
research
05/05/2021

Full-Sentence Models Perform Better in Simultaneous Translation Using the Information Enhanced Decoding Strategy

Simultaneous translation, which starts translating each sentence after r...

Please sign up or login with your details

Forgot password? Click here to reset