A Convolutional Neural Network for Modelling Sentences

作者: Nal Kalchbrenner , Edward Grefenstette , Phil Blunsom

DOI: 10.3115/V1/P14-1062

关键词:

摘要: The ability to accurately represent sentences is central language understanding. We describe a convolutional architecture dubbed the Dynamic Convolutional Neural Network (DCNN) that we adopt for semantic modelling of sentences. network uses k-Max Pooling, global pooling operation over linear sequences. handles input varying length and induces feature graph sentence capable explicitly capturing short long-range relations. does not rely on parse tree easily applicable any language. test DCNN in four experiments: small scale binary multi-class sentiment prediction, six-way question classification Twitter prediction by distant supervision. achieves excellent performance first three tasks greater than 25% error reduction last task with respect strongest baseline.

参考文章(35)
Dimitri Kartsaklis, Mehrnoosh Sadrzadeh, Prior Disambiguation of Word Tensors for Constructing Sentence Vectors empirical methods in natural language processing. pp. 1590- 1601 ,(2013)
Richard Socher, Andrew Y. Ng, Eric H. Huang, Christopher D. Manning, Jeffrey Pennington, Semi-Supervised Recursive Autoencoders for Predicting Sentiment Distributions empirical methods in natural language processing. pp. 151- 161 ,(2011)
Edward Grefenstette, Mehrnoosh Sadrzadeh, Experimental Support for a Categorical Compositional Distributional Model of Meaning empirical methods in natural language processing. pp. 1394- 1404 ,(2011)
Andreas Küchler, Christoph Goller, Inductive Learning in Symbolic Domains Using Structure-Driven Recurrent Neural Networks KI '96 Proceedings of the 20th Annual German Conference on Artificial Intelligence: Advances in Artificial Intelligence. pp. 183- 197 ,(1996) , 10.1007/3-540-61708-6_60
Luke S. Zettlemoyer, Michael Collins, Learning to map sentences to logical form: structured classification with probabilistic categorial grammars uncertainty in artificial intelligence. pp. 658- 666 ,(2005)
Alex Waibel, Kai-Fu Lee, Readings in speech recognition Morgan Kaufmann Publishers Inc.. ,(1990)
Katrin Erk, Vector Space Models of Word Meaning and Phrase Meaning: A Survey Language and Linguistics Compass. ,vol. 6, pp. 635- 653 ,(2012) , 10.1002/LNCO.362
Roberto Zamparelli, Marco Baroni, Nouns are Vectors, Adjectives are Matrices: Representing Adjective-Noun Constructions in Semantic Space empirical methods in natural language processing. pp. 1183- 1193 ,(2010)
Edward Thomas Grefenstette, Category-theoretic quantitative compositional distributional models of natural language semantics arXiv: Computation and Language. ,(2013)
Nal Kalchbrenner, Phil Blunsom, Recurrent Continuous Translation Models empirical methods in natural language processing. pp. 1700- 1709 ,(2013)