反向翻译back-translations

参考论文:

Rico Sennrich, Barry Haddow, and Alexandra Birch.2016. Edinburgh neural machine translation systems for wmt 16. arXiv preprint arXiv:1606.02891.

Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016a. Improving Neural Machine Translation Models with Monolingual Data. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL 2016), Berlin, Germany.

还是这个Sennrich,他在WMT16中说:We found that during decoding, the model would occasionally assign a high probability to
words based on the target context alone, ignoring the source sentence.

具体做法是 we experiment with training separate models that produce the target text from right-to-left (r2l), and re-scoring the nbest lists that are produced by the main (left-toright) models with these r2l models. Since the right-to-left model will see a complementary target context at each time step, we expect that the averaged probabilities will be more robust.

利用反向翻译模型对主模型翻译出的候选list进行了重打分。
 

在反向翻译时,为了增强模型的抗噪声能力,会人为加入噪声,常用的方法是drop out 

这个人用了一个简单粗暴的方法:In our English↔Romanian experiments, we drop out full words (both on the source and target side) with a probability of 0.1. For all other layers, the dropout probability is set to 0.2.

猜你喜欢

转载自blog.csdn.net/weixin_40240670/article/details/86552891