Using Self-Training to Improve Back-Translation in Low Resource Neural Machine Translation

作者: Bashir Shehu Galadanci , Idris Abdulmumin , Abubakar Isa

DOI:

关键词:

摘要: Improving neural machine translation (NMT) models using the back-translations of monolingual target data (synthetic parallel data) is currently state-of-the-art approach for training improved systems. The quality backward system - which trained on available and used back-translation has been shown in many studies to affect performance final NMT model. In low resource conditions, usually not enough train a model that can produce qualitative synthetic needed standard This work proposes self-training strategy where output improve itself through forward technique. technique was baseline IWSLT'14 English-German IWSLT'15 English-Vietnamese by 11.06 1.5 BLEUs respectively. generated out-performed another 2.7 BLEU.

参考文章(13)
Sepp Hochreiter, Jürgen Schmidhuber, Long short-term memory Neural Computation. ,vol. 9, pp. 1735- 1780 ,(1997) , 10.1162/NECO.1997.9.8.1735
Lucia Specia, Kashif Shah, Trevor Cohn, Jose G.C. de Souza, QuEst - A translation quality estimation framework meeting of the association for computational linguistics. pp. 79- 84 ,(2013)
Marcello Federico, Mauro Cettolo, Christian Girardi, WIT3: Web Inventory of Transcribed and Translated Talks Proceedings of the 16th Annual conference of the European Association for Machine Translation. pp. 261- 268 ,(2012)
Jiajun Zhang, Chengqing Zong, Exploiting Source-side Monolingual Data in Neural Machine Translation empirical methods in natural language processing. pp. 1535- 1545 ,(2016) , 10.18653/V1/D16-1160
Caglar Gulcehre, Orhan Firat, Kelvin Xu, Kyunghyun Cho, Yoshua Bengio, On integrating a language model into neural machine translation Computer Speech & Language. ,vol. 45, pp. 137- 148 ,(2017) , 10.1016/J.CSL.2017.01.014
Anna Currey, Antonio Valerio Miceli Barone, Kenneth Heafield, Copied Monolingual Data Improves Low-Resource Neural Machine Translation Proceedings of the Second Conference on Machine Translation. pp. 148- 156 ,(2017) , 10.18653/V1/W17-4715
Zhen Yang, Wei Chen, Feng Wang, Bo Xu, Effectively training neural machine translation models with monolingual data Neurocomputing. ,vol. 333, pp. 240- 247 ,(2019) , 10.1016/J.NEUCOM.2018.12.032
Rico Sennrich, Barry Haddow, Alexandra Birch, Improving Neural Machine Translation Models with Monolingual Data meeting of the association for computational linguistics. ,vol. 1, pp. 86- 96 ,(2016) , 10.18653/V1/P16-1009
Marc'Aurelio Ranzato, Michael Auli, Sumit Chopra, Wojciech Zaremba, Sequence Level Training with Recurrent Neural Networks international conference on learning representations. ,(2016)