Tencent AI Lab 官网
Target Foresight based Attention for Neural Machine
Abstract

In neural machine translation, an attention model is used to identify the aligned source words for a target word (target foresight word)in order to select translation context, but it does not make use of any information of this target foresight word at all.Previous work proposed an approach to improve the attention model by explicitly accessing this target foresight word and demonstrated the substantial gains in alignment task. However,this approach is useless in machine translation task on which the target foresight word is unavailable. In this paper, we propose a new attention model enhanced by the implicit information of target foresight word oriented to both alignment and translation tasks. Empirical experiments on Chinese-to-English and Japanese-to-English datasets show that the proposed attention model delivers significant improvements in terms of both alignment error rate and BLEU.


Venue
NAACL 2018
Publication Time
2018
Authors
Xintong Li,Lemao Liu,Zhaopeng Tu,Shuming Shi,Max Meng