A Generative Framework for Simultaneous Machine Translation

Yishu Miao, Phil Blunsom, Lucia Specia


Abstract
We propose a generative framework for simultaneous machine translation. Conventional approaches use a fixed number of source words to translate or learn dynamic policies for the number of source words by reinforcement learning. Here we formulate simultaneous translation as a structural sequence-to-sequence learning problem. A latent variable is introduced to model read or translate actions at every time step, which is then integrated out to consider all the possible translation policies. A re-parameterised Poisson prior is used to regularise the policies which allows the model to explicitly balance translation quality and latency. The experiments demonstrate the effectiveness and robustness of the generative framework, which achieves the best BLEU scores given different average translation latencies on benchmark datasets.
Anthology ID:
2021.emnlp-main.536
Volume:
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2021
Address:
Online and Punta Cana, Dominican Republic
Editors:
Marie-Francine Moens, Xuanjing Huang, Lucia Specia, Scott Wen-tau Yih
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6697–6706
Language:
URL:
https://aclanthology.org/2021.emnlp-main.536
DOI:
10.18653/v1/2021.emnlp-main.536
Bibkey:
Cite (ACL):
Yishu Miao, Phil Blunsom, and Lucia Specia. 2021. A Generative Framework for Simultaneous Machine Translation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6697–6706, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
Cite (Informal):
A Generative Framework for Simultaneous Machine Translation (Miao et al., EMNLP 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.emnlp-main.536.pdf
Video:
 https://aclanthology.org/2021.emnlp-main.536.mp4