论文标题

端到端的神经单词对齐优于giza ++

End-to-End Neural Word Alignment Outperforms GIZA++

论文作者

Zenkel, Thomas, Wuebker, Joern, DeNero, John

论文摘要

单词对齐曾经是自然语言处理中的核心无监督学习任务,因为它在训练统计机器翻译(MT)模型中的重要作用。尽管对训练神经MT模型不必要,但单词对准仍然在神经机器翻译的交互式应用中起重要作用,例如注释转移和词典注射。虽然统计MT方法已被具有卓越性能的神经方法取代,但二十岁的Giza ++工具包仍然是最新单词对准系统的关键组成部分。关于神经单词一致性的先前工作只能通过在训练过程中使用其输出来胜过Giza ++的表现。我们提出了第一个端到端神经单词对准方法,该方法在三个数据集上始终优于giza ++。我们的方法重新利用了一种经过培训的有监督翻译的变压器模型,以紧密整合并且不影响翻译质量的方式,也可以作为无监督的单词对齐模型。

Word alignment was once a core unsupervised learning task in natural language processing because of its essential role in training statistical machine translation (MT) models. Although unnecessary for training neural MT models, word alignment still plays an important role in interactive applications of neural machine translation, such as annotation transfer and lexicon injection. While statistical MT methods have been replaced by neural approaches with superior performance, the twenty-year-old GIZA++ toolkit remains a key component of state-of-the-art word alignment systems. Prior work on neural word alignment has only been able to outperform GIZA++ by using its output during training. We present the first end-to-end neural word alignment method that consistently outperforms GIZA++ on three data sets. Our approach repurposes a Transformer model trained for supervised translation to also serve as an unsupervised word alignment model in a manner that is tightly integrated and does not affect translation quality.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源