Relation Extraction with Matrix Factorization and Universal Schemas

作者: Benjamin M. Marlin , Sebastian Riedel , Andrew McCallum , Limin Yao

DOI:

关键词: Computer scienceData modelInformation retrievalMatrix decompositionSchema (psychology)Star schemaRelationship extractionTupleComputational linguistics

摘要: © 2013 Association for Computational Linguistics. Traditional relation extraction predicts relations within some fixed and finite target schema. Machine learning approaches to this task require either manual annotation or, in the case of distant supervision, existing structured sources same The need datasets can be avoided by using a universal schema: union all involved schemas (surface form predicates as OpenIE, preexisting databases). This schema has an almost unlimited set (due surface forms), supports integration with data (through types To populate database such we present matrix factorization models that learn latent feature vectors entity tuples relations. We show achieve substantially higher accuracy than traditional classification approach. More importantly, operating simultaneously on observed text pre-existing DBs Freebase, are able reason about unstructured mutually-supporting ways. By doing so our approach outperforms stateof- the-Art supervision.

参考文章(30)
Hristo Tanev, Ido Dagan, Idan Szpektor, Bonaventura Coppola, Scaling Web-based Acquisition of Entailment Relations. empirical methods in natural language processing. pp. 41- 48 ,(2004)
Sebastian Riedel, Andrew McCallum, Limin Yao, Probabilistic Databases of Universal Schema north american chapter of the association for computational linguistics. pp. 116- 121 ,(2012)
Sebastian Riedel, Aria Haghighi, Andrew McCallum, Limin Yao, Structured Relation Discovery using Generative Models empirical methods in natural language processing. pp. 1456- 1466 ,(2011)
Mihai Surdeanu, Ramesh Nallapati, Julie Tibshirani, Christopher D. Manning, Multi-instance Multi-label Learning for Relation Extraction empirical methods in natural language processing. pp. 455- 465 ,(2012)
Brian Murphy, Tom Mitchell, Partha Talukdar, Learning Effective and Interpretable Semantic Models using Non-Negative Sparse Embedding international conference on computational linguistics. pp. 1933- 1950 ,(2012)
Andrew Carlson, Justin Betteridge, Bryan Kisiel, Burr Settles, Estevam Hruschka, Tom Mitchell, None, Toward an architecture for never-ending language learning national conference on artificial intelligence. pp. 1306- 1313 ,(2010)
Timothy Chklovski, Rahul Bhagat, Eduard Hovy, Bonaventura Coppola, Patrick Pantel, ISP: Learning Inferential Selectional Preferences north american chapter of the association for computational linguistics. pp. 564- 571 ,(2007)
Hinrich Schütze, Christopher D. Manning, Prabhakar Raghavan, Introduction to Information Retrieval ,(2005)
Oren Etzioni, Stefan Schoenmackers, Daniel Weld, Jesse Davis, Learning First-Order Horn Clauses from Web Text empirical methods in natural language processing. pp. 1088- 1098 ,(2010)
Sebastian Riedel, Limin Yao, Andrew McCallum, Modeling relations and their mentions without labeled text european conference on machine learning. pp. 148- 163 ,(2010) , 10.1007/978-3-642-15939-8_10