Minimum Error Rate Training by Sampling the Translation Lattice

作者: Nicola Cancedda , Samidh Chatterjee

DOI:

关键词: Computer scienceWord error rateBLEUWork (physics)AlgorithmSampling (statistics)Machine translationLattice (module)Translation (geometry)Measure (mathematics)

摘要: Minimum Error Rate Training is the algorithm for log-linear model parameter training most used in state-of-the-art Statistical Machine Translation systems. In its original formulation, uses N-best lists output by decoder to grow Pool that shapes surface on which actual optimization performed. Recent work has been done extend use entire translation lattice built decoder, instead of lists. We propose here a third, intermediate way, consisting growing pool using samples randomly drawn from lattice. empirically measure systematic improvement BLEU scores compared lists, without suffering increase computational complexity associated with operating whole

参考文章(17)
Hideki Isozaki, Hajime Tsukada, Taro Watanabe, Jun Suzuki, Online Large-Margin Training for Statistical Machine Translation empirical methods in natural language processing. pp. 764- 773 ,(2007)
Kevin Knight, Decoding complexity in word-replacement translation models Computational Linguistics. ,vol. 25, pp. 607- 615 ,(1999)
Ultraconservative online algorithms for multiclass problems Journal of Machine Learning Research. ,vol. 3, pp. 951- 991 ,(2003) , 10.1162/JMLR.2003.3.4-5.951
Kishore Papineni, Salim Roukos, Todd Ward, Wei-Jing Zhu, BLEU Proceedings of the 40th Annual Meeting on Association for Computational Linguistics - ACL '02. pp. 311- 318 ,(2001) , 10.3115/1073083.1073135
Wolfgang Macherey, Franz Josef Och, Ignacio Thayer, Jakob Uszkoreit, Lattice-based minimum error rate training for statistical machine translation Proceedings of the Conference on Empirical Methods in Natural Language Processing - EMNLP '08. pp. 725- 734 ,(2008) , 10.3115/1613715.1613807
Richard Zens, Sasa Hasan, Hermann Ney, A Systematic Comparison of Training Criteria for Statistical Machine Translation empirical methods in natural language processing. pp. 524- 532 ,(2007)
David Chiang, Yuval Marton, Philip Resnik, Online large-margin training of syntactic and structural translation features Proceedings of the Conference on Empirical Methods in Natural Language Processing - EMNLP '08. pp. 224- 233 ,(2008) , 10.3115/1613715.1613747
Philipp Koehn, Richard Zens, Chris Dyer, Ondřej Bojar, Alexandra Constantin, Evan Herbst, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Moses: Open Source Toolkit for Statistical Machine Translation meeting of the association for computational linguistics. pp. 177- 180 ,(2007) , 10.3115/1557769.1557821
David A. Smith, Jason Eisner, Minimum Risk Annealing for Training Log-Linear Models meeting of the association for computational linguistics. pp. 787- 794 ,(2006) , 10.3115/1273073.1273174
Franz Josef Och, Minimum error rate training in statistical machine translation Proceedings of the 41st Annual Meeting on Association for Computational Linguistics - ACL '03. pp. 160- 167 ,(2003) , 10.3115/1075096.1075117