Bilingual continuous-space language model growing for statistical machine translation

作者: Rui Wang , Hai Zhao , Bao-Liang Lu , Masao Utiyama , Eiichiro Sumita

DOI: 10.1109/TASLP.2015.2425220

关键词:

摘要: Larger n-gram language models (LMs) perform better in statistical machine translation (SMT). However, the existing approaches have two main drawbacks for constructing larger LMs: 1) it is not convenient to obtain corpora same domain as bilingual parallel SMT; 2) most of previous studies focus on monolingual information from target only, and redundant n-grams been fully utilized SMT. Nowadays, continuous-space model (CSLM), especially neural network (NNLM), has shown great improvement estimation accuracies probabilities predicting words. these CSLM NNLM still consider only or require additional corpus. In this paper, we propose a novel based LM growing method. Compared approaches, proposed method enables us use corpus The results show that our new outperforms both SMT performance computational efficiency significantly.

参考文章(64)
Hai Zhao, Masao Utiyama, Eiichiro Sumita, Bao-Liang Lu, An empirical study on word segmentation for chinese machine translation international conference on computational linguistics. pp. 248- 263 ,(2013) , 10.1007/978-3-642-37256-8_21
Alexandre Allauzen, Guillaume Wisniewski, Hai Son Le, François Yvon, Training Continuous Space Language Models: Some Practical Issues empirical methods in natural language processing. pp. 778- 788 ,(2010)
Victor Abrash, Andreas Stolcke, Wen Wang, Jing Zheng, SRILM at Sixteen: Update and Outlook IEEE SPS. ,(2011)
Joel Martin, Roland Kuhn, George Foster, Howard Johnson, Improving Translation Quality by Discarding Most of the Phrasetable empirical methods in natural language processing. pp. 967- 975 ,(2007)
Tomas Mikolov, Martin Karafiát, Sanjeev Khudanpur, Jan Cernocký, Lukás Burget, Recurrent neural network based language model conference of the international speech communication association. pp. 1045- 1048 ,(2010)
Philipp Koehn, Statistical Significance Tests for Machine Translation Evaluation. empirical methods in natural language processing. pp. 388- 395 ,(2004)
Yinggong Zhao, Ashish Vaswani, David Chiang, Victoria Fossum, Decoding with Large-Scale Neural Language Models Improves Translation empirical methods in natural language processing. pp. 1387- 1392 ,(2013)
Andreas Stolcke, SRILM – An Extensible Language Modeling Toolkit conference of the international speech communication association. ,(2002)
Nal Kalchbrenner, Phil Blunsom, Recurrent Continuous Translation Models empirical methods in natural language processing. pp. 1700- 1709 ,(2013)
Andreas Stolcke, Entropy-based Pruning of Backoff Language Models arXiv: Computation and Language. ,(2000)