NRE: Neural Relation Extraction.
Contributed by Tianyu Gao and Xu Han.
We released OpenNRE, an open-source framework for neural relation extraction. This repository provides several relation extraction methods and an easy-to-use training and testing framework.
A Survey of Deep Learning Methods for Relation Extraction.
Shantanu Kumar.
2017.
paper
Relation Extraction : A Survey.
Sachin Pawara,b, Girish K. Palshikara, Pushpak Bhattacharyyab.
2017.
paper
ACE 2005 Dataset link
SemEval-2010 Task 8 Dataset link
FewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation
Xu Han, Hao Zhu, Pengfei Yu, Ziyun Wang, Yuan Yao, Zhiyuan Liu, Maosong Sun
EMNLP 2018.
paper
We present a Few-Shot Relation Classification Dataset (FewRel), consisting of 70,000 sentences on 100 relations derived from Wikipedia and annotated by crowdworkers.
Word2vec link
GloVe link
SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals.
Iris Hendrickx , Su Nam Kim, Zornitsa Kozareva, Preslav Nakov, Diarmuid O ́ Se ́aghdha, Sebastian Pado ́, Marco Pennacchiotti, Lorenza Romano, Stan Szpakowicz.
Workshop on Semantic Evaluations, ACL 2009
paper
This leads us to introduce a new task, which will be part of SemEval-2010: multi-way classification of mutually exclusive semantic relations between pairs of common nominals.
Learning to Extract Relations from the Web using Minimal Supervision.
Razvan C. Bunescu, Department of Computer Sciences.
ACL 2007.
paper
We present a new approach to relation extraction that requires only a handful of training examples. Given a few pairs of named entities known to exhibit or not exhibit a particular relation, bags of sentences containing the pairs are extracted from the web.
Distant Supervision for Relation Extraction without Labeled Data.
Mike Mintz, Steven Bills, Rion Snow, Dan Jurafsky.
ACL-IJCNLP 2009.
paper
Our experiments use Freebase, a large semantic database of several thousand relations, to provide distant supervision.
Modeling Relations and Their Mentions without Labeled Text.
Sebastian Riedel, Limin Yao, Andrew McCallum.
ECML 2010.
paper
We present a novel approach to distant supervision that can alleviate this problem based on the following two ideas: First, we use a factor graph to explicitly model the decision whether two entities are related, and the decision whether this relation is mentioned in a given sentence; second, we apply constraint-driven semi-supervision to train this model without any knowledge about which sentences express the relations in our training KB.
Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations.
Raphael Hoffmann, Congle Zhang, Xiao Ling, Luke Zettlemoyer, Daniel S. Weld.
ACL-HLT 2011.
paper
This paper presents a novel approach for multi-instance learning with overlapping re- lations that combines a sentence-level extrac- tion model with a simple, corpus-level compo- nent for aggregating the individual facts.
Distributed Representations of Words and Phrases and their Compositionality.
Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, Jeffrey Dean.
NIPS 2013.
paper
In this paper we present several extensions that improve both the quality of the vectors and the training speed. By subsampling of the frequent words we obtain significant speedup and also learn more regular word representations. We also describe a simple alternative to the hierarchical softmax called negative sampling.
GloVe: Global Vectors for Word Representation.
Jeffrey Pennington, Richard Socher, Christopher D. Manning.
EMNLP 2014.
paper
The result is a new global log-bilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods.
Semantic Compositionality through Recursive Matrix-Vector Spaces.
Richard Socher, Brody Huval, Christopher D. Manning, Andrew Y. Ng.
EMNLP-CoNLL 2012.
paper
We introduce a recursive neural network (RNN) model that learns compositional vector representations for phrases and sentences of arbitrary syntactic type and length.
Convolution Neural Network for Relation Extraction
Chunyang Liu, Wenbo Sun, Wenhan Chao, Wanxiang Che.
ADMA 2013
paper
In this paper, we propose a novel convolution network, incorporating lexical features, applied to Relation Extraction.
Relation Classification via Convolutional Deep Neural Network.
Daojian Zeng, Kang Liu, Siwei Lai, Guangyou Zhou, Jun Zhao.
COLING 2014.
paper
We exploit a convolutional deep neural network (DNN) to extract lexical and sentence level features. Our method takes all of the word tokens as input without complicated pre-processing.
Classifying Relations by Ranking with Convolutional Neural Networks.
C´ıcero Nogueira dos Santos, Bing Xiang, Bowen Zhou.
ACL 2015.
paper
In this work we tackle the relation classification task using a convolutional neural network that performs classification by ranking (CR-CNN).
Relation Extraction: Perspective from Convolutional Neural Networks.
Thien Huu Nguyen, Ralph Grishman.
NAACL-HLT 2015
paper
Our model takes advantages of multiple window sizes for filters and pre-trained word embeddings as an initializer on a non-static architecture to improve the performance. We emphasize the relation extraction problem with an unbalanced corpus.
End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures.
Makoto Miwa, Mohit Bansal.
ACL 2016.
paper
Our recurrent neural network based model captures both word sequence and dependency tree substructure information by stacking bidirectional tree-structured LSTM-RNNs on bidirectional sequential LSTM-RNNs… We further encourage detection of entities during training and use of entity information in relation extraction via entity pre-training and scheduled sampling.
A Walk-based Model on Entity Graphs for Relation Extraction.
Fenia Christopoulou, Makoto Miwa, Sophia Ananiadou.
ACL 2018.
paper
We present a novel graph-based neural network model for relation extraction. Our model treats multiple pairs in a sentence simultaneously and considers interactions among them. All the entities in a sentence are placed as nodes in a fully-connected graph structure.
Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks.
Daojian Zeng, Kang Liu, Yubo Chen, Jun Zhao.
EMNLP 2015.
paper
We propose a novel model dubbed the Piecewise Convolutional Neural Networks (PCNNs) with multi-instance learning to address these two problems.
Neural Relation Extraction with Selective Attention over Instances.
Yankai Lin, Shiqi Shen, Zhiyuan Liu, Huanbo Luan, Maosong Sun.
ACL 2016.
paper
Distant supervision inevitably accompanies with the wrong labelling problem, and these noisy data will substantially hurt the performance of relation extraction. To alleviate this issue, we propose a sentence-level attention-based model for relation extraction.
Relation Extraction with Multi-instance Multi-label Convolutional Neural Networks.
Xiaotian Jiang, Quan Wang, Peng Li, Bin Wang.
COLING 2016.
paper
In this paper, we propose a multi-instance multi-label convolutional neural network for distantly supervised RE. It first relaxes the expressed-at-least-once assumption, and employs cross-sentence max-pooling so as to enable information sharing across different sentences.
Adversarial Training for Relation Extraction.
Yi Wu, David Bamman, Stuart Russell.
EMNLP 2017.
paper
Adversarial training is a mean of regularizing classification algorithms by generating adversarial noise to the training data. We apply adversarial training in relation extraction within the multi-instance multi-label learning framework.
A Soft-label Method for Noise-tolerant Distantly Supervised Relation Extraction.
Tianyu Liu, Kexiang Wang, Baobao Chang, Zhifang Sui.
EMNLP 2017.
paper
We introduce an entity-pair level denoise method which exploits semantic information from correctly labeled entity pairs to correct wrong labels dynamically during training.
DSGAN: Generative Adversarial Training for Distant Supervision Relation Extraction.
Pengda Qin, Weiran Xu, William Yang Wang.
paper
We introduce an adversarial learning framework, which we named DSGAN, to learn a sentence-level true-positive generator. Inspired by Generative Adversarial Networks, we regard the positive samples generated by the generator as the negative samples to train the discriminator.
Reinforcement Learning for Relation Classification from Noisy Data.
Jun Feng, Minlie Huang, Li Zhao, Yang Yang, Xiaoyan Zhu.
AAAI 2018.
paper
We propose a novel model for relation classification at the sentence level from noisy data. The model has two modules: an instance selector and a relation classifier. The instance selector chooses high-quality sentences with reinforcement learning and feeds the selected sentences into the relation classifier, and the relation classifier makes sentence-level prediction and provides rewards to the instance selector.
Robust Distant Supervision Relation Extraction via Deep Reinforcement Learning.
Pengda Qin, Weiran Xu, William Yang Wang.
2018.
paper
We explore a deep reinforcement learning strategy to generate the false-positive indicator, where we automatically recognize false positives for each relation type without any supervised information.
Neural Knowledge Acquisition via Mutual Attention between Knowledge Graph and Text
Xu Han, Zhiyuan Liu, Maosong Sun.
AAAI 2018.
paper
We propose a general joint representation learning framework for knowledge acquisition (KA) on two tasks, knowledge graph completion (KGC) and relation extraction (RE) from text. We propose an effective mutual attention between KGs and text. The recip- rocal attention mechanism enables us to highlight important features and perform better KGC and RE.
Hierarchical Relation Extraction with Coarse-to-Fine Grained Attention
Xu Han, Pengfei Yu, Zhiyuan Liu, Maosong Sun, Peng Li.
EMNLP 2018.
paper
We aim to incorporate the hierarchical information of relations for distantly supervised relation extraction and propose a novel hierarchical attention scheme. The multiple layers of our hierarchical attention scheme provide coarse-to-fine granularity to better identify valid instances, which is especially effective for extracting those long-tail relations.
Incorporating Relation Paths in Neural Relation Extraction
Wenyuan Zeng, Yankai Lin, Zhiyuan Liu, Maosong Sun.
EMNLP 2017.
paper
We build inference chains between two target entities via intermediate entities, and propose a path-based neural relation extraction model to encode the relational semantics from both direct sentences and inference chains.
RESIDE: Improving Distantly-Supervised Neural Relation Extractionusing Side Information
Shikhar Vashishth, Rishabh Joshi, Sai Suman Prayaga, Chiranjib Bhattacharyya, Partha Talukdar.
EMNLP 2018.
paper
In this paper, we propose RESIDE, a distantly-supervised neural relation extraction method which utilizes additional side information from KBs for improved relation extraction. It uses entity type and relation alias information for imposing soft constraints while predicting relations.