Viterbi Decoding of Directed Acyclic Transformer for Non-Autoregressive Machine Translation

10/11/2022
by   Chenze Shao, et al.
0

Non-autoregressive models achieve significant decoding speedup in neural machine translation but lack the ability to capture sequential dependency. Directed Acyclic Transformer (DA-Transformer) was recently proposed to model sequential dependency with a directed acyclic graph. Consequently, it has to apply a sequential decision process at inference time, which harms the global translation accuracy. In this paper, we present a Viterbi decoding framework for DA-Transformer, which guarantees to find the joint optimal solution for the translation and decoding path under any length constraint. Experimental results demonstrate that our approach consistently improves the performance of DA-Transformer while maintaining a similar decoding speedup.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2019

Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation

Non-Autoregressive Transformer (NAT) aims to accelerate the Transformer ...
research
11/06/2019

Guiding Non-Autoregressive Neural Machine Translation Decoding with Reordering Information

Non-autoregressive neural machine translation (NAT) generates each targe...
research
03/14/2023

RenewNAT: Renewing Potential Translation for Non-Autoregressive Transformer

Non-autoregressive neural machine translation (NAT) models are proposed ...
research
06/05/2019

Imitation Learning for Non-Autoregressive Neural Machine Translation

Non-autoregressive translation models (NAT) have achieved impressive inf...
research
03/12/2023

Fuzzy Alignments in Directed Acyclic Graph for Non-Autoregressive Machine Translation

Non-autoregressive translation (NAT) reduces the decoding latency but su...
research
05/17/2023

Accelerating Transformer Inference for Translation via Parallel Decoding

Autoregressive decoding limits the efficiency of transformers for Machin...
research
05/29/2019

A Generalized Framework of Sequence Generation with Application to Undirected Sequence Models

Undirected neural sequence models such as BERT have received renewed int...

Please sign up or login with your details

Forgot password? Click here to reset