Iterative Self-Learning for Enhanced Back-Translation in Low Resource Neural Machine Translation.

作者: Bashir Shehu Galadanci , Idris Abdulmumin , Ismaila Idris Sinan

DOI:

关键词:

摘要: Many language pairs are low resource - the amount and/or quality of parallel data is not sufficient to train a neural machine translation (NMT) model which can reach an acceptable standard accuracy. works have explored use easier-to-get monolingual improve performance models in this category languages and even high languages. The most successful such back-translation using translations target increase training data. backward trained on available has been shown determine approach. approaches especially where model. Among self-learning iterative back-translation. These methods were perform better than This work presents self-training approach as improvement over further enhance Over several iterations, synthetic generated by used its through forward translation. Experiments that method outperforms both IWSLT'14 English German NMT. While also back-translation, though slightly, number required be reduced exactly iterations.

参考文章(39)
Diederik P. Kingma, Jimmy Ba, Adam: A Method for Stochastic Optimization arXiv: Learning. ,(2014)
Rico Sennrich, Alexandra Birch, Barry Haddow, Neural Machine Translation of Rare Words with Subword Units arXiv: Computation and Language. ,(2015)
Thang Luong, Hieu Pham, Christopher D. Manning, Effective Approaches to Attention-based Neural Machine Translation empirical methods in natural language processing. pp. 1412- 1421 ,(2015) , 10.18653/V1/D15-1166
Sepp Hochreiter, Jürgen Schmidhuber, Long short-term memory Neural Computation. ,vol. 9, pp. 1735- 1780 ,(1997) , 10.1162/NECO.1997.9.8.1735
Alon Lavie, Michael J. Denkowski, The Meteor metric for automatic evaluation of machine translation Machine Translation. ,vol. 23, pp. 105- 115 ,(2009) , 10.1007/S10590-009-9059-4
Kishore Papineni, Salim Roukos, Todd Ward, Wei-Jing Zhu, BLEU Proceedings of the 40th Annual Meeting on Association for Computational Linguistics - ACL '02. pp. 311- 318 ,(2001) , 10.3115/1073083.1073135
Yoshua Bengio, Kyunghyun Cho, Dzmitry Bahdanau, Neural Machine Translation by Jointly Learning to Align and Translate arXiv: Computation and Language. ,(2014)
Lucia Specia, Kashif Shah, Trevor Cohn, Jose G.C. de Souza, QuEst - A translation quality estimation framework meeting of the association for computational linguistics. pp. 79- 84 ,(2013)
Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, Manjunath Kudlur, Josh Levenberg, Rajat Monga, Sherry Moore, Derek G Murray, Benoit Steiner, Paul Tucker, Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan Yu, Xiaoqiang Zheng, None, TensorFlow: a system for large-scale machine learning operating systems design and implementation. pp. 265- 283 ,(2016) , 10.5555/3026877.3026899
Jiajun Zhang, Chengqing Zong, Exploiting Source-side Monolingual Data in Neural Machine Translation empirical methods in natural language processing. pp. 1535- 1545 ,(2016) , 10.18653/V1/D16-1160