Morphosyntactic Tagging with a Meta-BiLSTM Model over Context Sensitive Token Encodings

作者: Bernd Bohnet , Emily Pitler , Goncalo Simoes , Daniel Andor , Joshua Maynez

DOI:

关键词:

摘要: The rise of neural networks, and particularly recurrent has produced significant advances in part-of-speech tagging accuracy. One characteristic common among these models is the presence rich initial word encodings. These encodings typically are composed a character-based representation with learned pre-trained embeddings. However, do not consider context wider than single it only through subsequent layers that or sub-word information interacts. In this paper, we investigate use networks sentence-level for character word-based representations. particular show optimal results obtained by integrating sensitive representations synchronized training meta-model learns to combine their states. We present on morphological state-of-the-art performance number languages.

参考文章(16)
Wang Ling, Chris Dyer, Alan W Black, Isabel Trancoso, Ramon Fermandez, Silvio Amir, Luis Marujo, Tiago Luis, Finding Function in Form: Compositional Character Models for Open Vocabulary Word Representation empirical methods in natural language processing. pp. 1520- 1530 ,(2015) , 10.18653/V1/D15-1176
Bianca Zadrozny, Cicero Dos Santos, Learning Character-level Representations for Part-of-Speech Tagging international conference on machine learning. pp. 1818- 1826 ,(2014)
Ronan Collobert, Jason Weston, A unified architecture for natural language processing Proceedings of the 25th international conference on Machine learning - ICML '08. pp. 160- 167 ,(2008) , 10.1145/1390156.1390177
M. Schuster, K.K. Paliwal, Bidirectional recurrent neural networks IEEE Transactions on Signal Processing. ,vol. 45, pp. 2673- 2681 ,(1997) , 10.1109/78.650093
Anders Sogaard, Semi-supervised condensed nearest neighbor for part-of-speech tagging meeting of the association for computational linguistics. pp. 48- 52 ,(2011)
Chris Alberti, David Weiss, Greg Coppola, Slav Petrov, Improved Transition-Based Parsing and Tagging with Neural Networks empirical methods in natural language processing. pp. 1354- 1359 ,(2015) , 10.18653/V1/D15-1159
Danqi Chen, Christopher Manning, A Fast and Accurate Dependency Parser using Neural Networks Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). pp. 740- 750 ,(2014) , 10.3115/V1/D14-1082
Aliaksei Severyn, Michael Collins, Chris Alberti, Kuzman Ganchev, Alessandro Presta, Slav Petrov, Daniel Andor, David Weiss, Globally Normalized Transition-Based Neural Networks arXiv: Computation and Language. ,(2016)
Kris Cao, Marek Rei, A Joint Model for Word Embedding and Word Morphology meeting of the association for computational linguistics. pp. 18- 26 ,(2016) , 10.18653/V1/W16-1603
Jinho D. Choi, Dynamic Feature Induction: The Last Gist to the State-of-the-Art north american chapter of the association for computational linguistics. pp. 271- 281 ,(2016) , 10.18653/V1/N16-1031