Adversarial Learning for Distant Supervised Relation Extraction

2018-06-01 11:12DaojianZengYuanDaiFengLiSimonSherrattandJinWang
Computers Materials&Continua 2018年4期

Daojian Zeng , Yuan Dai , Feng Li , R. Simon Sherratt and Jin Wang,

1 Introduction

There have been many methods proposed for relation extraction. In these methods, the supervised paradigm has been shown to be effective and yields relatively high performance [Kambhatla (2004); Zhou, Su, Zhang et al. (2005)]. However, a large labeled training data is often required for supervision, and manually annotating large labeled training data is a time-consuming and labor-intensive task. In addition, since the manually labeled training data is often domain dependent, the model tends to be biased toward a specific domain.

Figure 1: An training example

To address the shortcomings of supervised paradigm, distantly supervised paradigm[Mintz, Bills, Snow et al. (2009)] is proposed to automatically generate training data.After obtaining the distant supervised labeled data, traditional methods sometimes applied supervised models to elaborately handcrafted features [Mintz, Bills, Snow et al.(2009); Riedel, Yao and McCallum (2010); Hoffmann, Zhang, Ling et al. (2011);Surdeanu, Tibshirani, Nallapati et al. (2012)]. These features are often derived from offthe-shelf Natural Language Processing (NLP) tools, which inevitably have errors, and have negative impact on the classification accuracy.

With the recent revival of interest in neural networks, many researchers have investigated the possibility of using neural networks to automatically learn features for relation classification [Zeng, Liu, Lai et al. (2014); Zeng, Liu, Chen et al. (2015); Jiang, Wang, Li et al. (2016); Lin, Shen, Liu et al. (2016)]. These approaches generally use a softmax classifier with cross-entropy loss and inevitably bring the noise of artificial class NA into classification process. To address the shortcoming, the classifier with ranking loss is employed to relation extraction [Zeng, Zeng and Dai (2017)]. Generally, the ranking loss function needs a negative class to train the model. Randomly selecting a relation or selecting the highest score among all incorrect relations are two common methods for generating negative class.

Unfortunately, these approaches are not ideal, because the sampled relation could be completely unrelated to the two given entities, so the majority of the generated negative class can be easily discriminated from positive class. Thus the quality of selected negative class is often poor and will contribute little towards the training. Cai et al. [Cai and Wang (2017)] and [Jia and Liang (2017)] have found that the quality of the negative class (pattern) is very important in the discriminative architecture. For example, the positive relation between Obama and Hawaii is /people/person/place_of birth in Fig. 1(ID 1). Obviously, there can be no /business/company/founders relation between a Person and a Location and it is a poor negative relation (ID 3). Accordingly, the birthplace of a person is probably also the place where he lives. Thus, /people/person/place_lived is the high-quality one that can be used to improve the model’s discrimination (ID 2).

In this paper, we provide a generic solution to improve the training of ranking based DSRE. Inspired by generative adversarial networks (GANs) [Goodfellow, Pouget-Abadie,Mirza et al. (2014)], we propose a novel adversarial learning framework for this task and use a neural network as the negative label generator to assist the training of our desired model, which acts as the discriminator in GANs. More specifically, we consider a two layers fully connected neural network as the generator to supply better quality negative labels at first. Then we adopt the PCNNs as the discriminator to classify the final relation.Through the alternating optimization of generator and discriminator, the generator is learning to produce more and more discriminable negative classes and the discriminator has to become better as well. Since the generator has a discrete generation step, we cannot directly use the gradient-based approach to back propagate the errors. We then consider a one-step reinforcement learning setting and use a REINFORCE method to achieve this goal [Watkins (1992)].

In sum, the main contributions of this paper lie in three folds:

● We combine GAN with the ranking-based approach and propose a new paradigm for DSRE. To our best knowledge, this is the first attempt to consider adversarial learning for this task.

● We prove that the generator can consistently provide high-quality negative classes which is crucial for the discriminator to improve DSRE.

● Empirically, we perform experiments on a widely used dataset and verify the adversarial learning approach. Experimental results show that our approach obtains state-of-the-art performance on the dataset.

2 Related work

2.1 Relation extraction

Relation extraction is one of the most important topics in NLP. Supervised approaches are the most commonly used methods for relation extraction and yield relatively high performance [Bunescu and Mooney (2005); Zelenko, Aone and Richardella (2003); Zhou,Su, Zhang et al. (2005)]. In the supervised paradigm, relation extraction is considered to be a multi-class classification problem and may suffer from a lack of labeled data for training. To address this issue, Mintz et al. [Mintz, Bills, Snow et al. (2009)] adopts Freebase to perform distant supervision. The algorithm for training data generation is sometimes faced with the wrong label problem. To address this shortcoming, Riedel et al.[Riedel, Yao and McCallum (2010); Hoffmann, Zhang, Ling et al. (2011); Surdeanu,Tibshirani, Nallapati et al. (2012)] develop the relaxed distant supervision assumption for multi-instance learning. Nguyen et al. [Nguyen and Moschitti (2011)] utilize relation definitions and Wikipedia documents to improve their systems.

The methods mentioned above have been shown to be effective for DSRE. However,their performance depends strongly on the quality of the designed features. Recently,deep learning has made great strides in many tasks [Gurusamy and Subramaniam (2017);Yuan, Li, Wu et al. (2017)]. Many researchers attempt to use deep neural network to automatically learning feature for DSRE. Zeng et al. [Zeng, Liu, Lai et al. (2014)] adopts CNNs to embed the semantics of the sentences. Moreover, Santos et al. [Santos, Xiang and Zhou (2015)] proposes a pairwise ranking loss function in the CNNs to reduce the impact of artificial class. These methods build classifier based on sentence-level annotated data, which cannot be directly applied for DSRE since multiple sentences corresponding to a fact may be achieved in the data generating procedure. Therefore,Zeng et al. [Zeng, Liu, Chen et al. (2015)] incorporate multi-instance learning with neural network model, which can build relation extractor based on distant supervision data.Although the method achieves significant improvement in relation extraction, it only selects the most likely sentence for each entity pair in their multi-instance learning paradigm. To address this issue, Lin et al. [Lin, Shen, Liu et al. (2016)] propose sentence level attention over multiple instances in order to utilize all informative sentences. Jiang et al. [Jiang, Wang, Li et al. (2016)] employ cross-sentence max-pooling to select features across different instances and then aggregates the most significant features for each entity pair.

The aforementioned works, especially neural networks, have greatly promoted the development of relation extraction. However, these works do not pay attention to the noise of artificial class NA, which are unfortunately very common in DSRE. Zeng et al.[Zeng, Zeng and Dai (2017)] proposed ranking loss and cost-sensitive to address the noise of NA. They select the highest score among all incorrect relations as the negative label. This approach is not ideal, because the quality of the selected label is often poor. In this paper, we propose a novel pair-wise ranking loss whose negative samples are provided by a generator of the GAN.

2.2 Generative adversarial networks

GANs [Goodfellow, Pouget-Abadie, Mirza et al. (2014)] was originally proposed for generating samples in a continuous space such as images. A GAN consists of two parts,the generator, and the discriminator. The generator accepts a noise input and outputs an image. The discriminator is a classifier which classifies images as “true” (from the ground truth set) or “fake” (generated by the generator). When training a GAN, the generator and the discriminator play a minimax game, in which the generator tries to generate “real” images to deceive the discriminator, and the discriminator tries to tell them apart from ground truth images. GANs are also capable of generating samples satisfying certain requirements, such as conditional GAN [Mirza and Osindero (2014)]. It is not possible to use GANs in its original form for generating discrete samples like natural language sentences or knowledge graph triples, because the discrete sampling step prevents gradients from propagating back to the generator. SeqGAN [Yu, Zhang, Wang et al. (2017)] is one of the first successful solutions to this problem by using reinforcement learning, which trains the generator using policy gradient. Likewise, our framework relies on policy gradient to train the generator which provides discrete negative labels.

3 Task definition

DSRE is usually considered as a multi-instance learning problem. In multi-instance learning paradigm, all sentences labeled by a relation triplet constitute a bag and each sentence is called an instance.

Suppose that there arein the training set and that the-th bagcontainsinstances.The objective of multi-instance learning is to predict the labels of the unseen bags. We need to learn a relation extractor based on the training data and then use it to predict relations for the test set.

Figure 2: GAN-based framework. (a) Generator: calculating a probability distribution over a set of candidate negative relations, then sample one relation from the distribution as the output. (b) Discriminator: receiving the generated negative triple as well as the ground truth triple (in the hexagonal box), and calculating their scores. Generator maximizes the score of the generated negative class by policy gradient, and discriminator minimizes the marginal loss between positive and negative class by gradient descent

Specifically, for a bagin training set, we need to extract features from the bag (from one or several valid instances) and then use them to train a classifier.

4 Methodology

Fig. 2 shows the neural network architecture used in this work. It consists of two parts:Discriminator Network and Generator Network. In this section, we first introduce the discriminator network and generator network used in this paper. Then, we define the objective function for discriminator and generator respectively and explain how to alternatively train the proposed model.

4.1 Discriminator network

The discriminator in our model is similar to [Zeng, Zeng and Dai (2017)] which is shown in Fig. 3. Different from their model that selects the highest score among all incorrect relations as a negative class, our model uses the negative label generated by the generator which will be described in detail in Section 4.2. In this section, we will first give a detailed description of the discriminator.

4.1.1 Vector representation

The inputs of our discriminator are raw word tokens. When using neural networks, we typically transform word tokens into low-dimensional vectors. In this paper, the “word token” refers to word and entity. In the following, we do not distinguish them and call them “word”. In our method, each input word token is transformed into a vector by looking up pre-trained word embeddings. Moreover, we use Position Features (PFs)[Zeng, Liu, Lai et al. (2014); Zeng, Liu, Chen et al. (2015)] to specify entity pairs, which are also transformed into vectors by looking up position embeddings.

Figure 3: Discriminator network

4.1.2 Word embeddings

Word embeddings are distributed representations of words that map each word in a text to a ‘k’-dimensional real-valued vector. They have recently been shown to capture both semantic and syntactic information about words very well, setting performance records in several word similarity tasks [Mikolov, Chen, Corrado et al. (2013); Pennington, Socher and Manning (2014)]. Using word embeddings that have been trained a priori has become common practice for enhancing many other NLP tasks [Huang, Ahuja, Downey et al.(2014)]. In the past years, many methods for training word embeddings have been proposed [Bengio, Ducharme, Vincent et al. (2003); Collobert, Weston, Bottou et al.(2011); Mikolov, Chen, Corrado et al. (2013)]. We employ the method [Mikolov, Chen,Corrado et al. (2013)] to train word embeddings and denote it by E.

4.1.3 Position embeddings

Zeng et al. [Zeng, Liu, Chen et al. (2015)] have shown the importance of PFs in relation extraction. Similar to their works, we use PFs to specify entity pairs. A PF is defined as the combination of the relative distances from the current word to entityand entity.We randomly initialize two position embedding matricesand transform the relative distances into vectors by looking them up.

We concatenate the word representation and position representation as the input of the network (shown in Fig. 3(a)). Assume that the size of word representation iskwand that of position representation is kd, then the size of a word vector is

4.1.4 Convolution

Assume thatand, then the convolution of A and B is defined as

We denote the input sentence by, whereis the-th word, and useto represent its vector. We useto represent the matrix concatenated by sequencedenotes the horizontal concatenation ofx1andx2).We denote the length of filter by(Fig. 3(a) shows an example of= 3), then the weight matrix of the filter is. Then the convolution operation between the filter and sentenceresults in another vector

where

In experiments, we usen( n>1)filters (or feature maps) to capture different features of an instance. Therefore, we also needweight matrices, so that all the convolution operations can be expressed by

where1≤i≤n and. Through the convolution layer, we obtain the results vectors

4.1.5 Piecewise max pooling

In order to capture the structural information and fine-grained features, PCNNs divides an instance into three segments according to the given entity pair (two entities cut the sentence into three parts) and do max-pooling operation on each segment. For the result vector ciof convolution operations, it can be divided into three partsThen piecewise max-pooling procedure iswhereand

After that, we can concatenate all the vectorsto obtain vector. Fig. 3(a) displays an example of, in which the gray circles are the positions of entities. Finally, we compute the feature vectorfor sentence S .

4.1.6 Classifier

To compute the score for each relation, the feature vector of each instance is fed into a pair-wise ranking loss based classifier. Given the distributed vector representation of an instance b , the network computes the score for a class label tiby using the dot product:

where w∈R1×3nis the class embedding for class label t. All the class embeddings

tiiconstitute the class embedding matrixwhose rows encode the distributed vector representations of the different class labels. Tis equal to the number of possible relation types for the relation extraction system. Note that the number of dimensions in each class embedding must be equal to the size of the distributed vector representation of the input bag3n . The class embedding matrix WTis a parameter to be learned by the network.

4.1.7 Instance selection

Distant supervised relation extraction suffers from wrong label problem [Riedel, Yao and McCallum (2010)]. The core problem that needs to be solved in the multi-instance learning is to get the corresponding bag feature vector from all the instance feature vectors in the bag. In fact, the problem is the instance selection strategy. We employ an instance selection strategy borrowed from Zeng et al. [Zeng, Liu, Chen et al. (2015)].Different from [Zeng, Liu, Chen et al. (2015)], we randomly select an instance from the bag with NA label since our model do not give score for NAclass (see Section 4.1.8.).In addition, we choose the instance which has the highest score for the bag label except forNA. The scores are computed using Eq. (3). Therefore, our instance selection strategy will not be disturbed by the noise inNA. Assume that there is a bagthat contains qiinstances with feature vectorsand the bag label is riThe j -th instanceis selected and thej is constrained as follows:

whereis computed using Eq. (3).

4.1.8 Pair-wise ranking loss

The cross-entropy loss brings the noise of artificial class into the classification process.This phenomenon is mainly due to the noise of artificial classNA . To address this shortcoming, we propose a new pairwise ranking loss instead of cross-entropy which is often used for softmax classifier.

In our model, the network can be stated as a tuple. Assume that there are N bags in training set, and their labels are relationsAfter the instance selection, we get a representative instance and its corresponding feature vector is considered as the bag feature vector. The input for each iteration round is a bag feature vector and the class label. In the pairwise loss, the loss function is defined on the basis of pairs of objects whose labels are different. We can get the loss function by selecting a class label that is different from the input one. In this work, we use the negative samples generated by the generator. For example, when the-th bag with ground truth labelis fed into the network, we will get a negative classwhich provided by the generator. The pair-wise ranking loss function is defined as follows:

whereis a scaling factor that magnifies the difference between the scores. m+and m-are the margin for correct and incorrect class,respectively. Since it is very difficult to learn patterns for the artificial class NA , softmax classifier often brings noise into the classification process of the natural classes. By using a ranking classifier, we can avoid explicitly leaning patterns for the artificial class. We omit the artificial classNA by setting the first term in the right side of Eq. (5) to zero and do not learn the class embedding forNA. Thus, our model does not give a score for the artificial classNA and the noise inNAis alleviated. At prediction time, an instance is classified asNAonly if all actual classes have negative scores. A bag is positively labeled if and only if the output of the network on at least one of its instances is assigned a positive label and we choose the class which has the highest score.

4.2 Generator network

The aim of our generator is to provide the discriminator with high-quality negative classes to improve DSRE. In this paper, we devise a two layers fully-connected neural network as the generator. As shown in Fig. 2 (a), the input of the generator is a vector v that combines the embeddings of two entities (and) and the embedding of the ground truth relationrt. Nonlinear function ReLU is added after the first layer. The output of the network is as follows:

The softmax function is added to the second layer because they can adequately model the“sampling from a probability distribution process” of discrete GANs. The probability distribution of the relation set ℜis defined as:

Finally, the generator samples one relation from the distributionas the output.

4.3 Generative adversarial training

Intuitively, the discriminator should assign a relatively large score for positive class and small score for the negative class. The objective of the discriminator can be formulated as minimizing the following ranking loss function:

The only difference between this loss function and Eq. (5) is that it uses negative class from the generator. In order to encourage the generator to generate useful negative classes, the objective of the generator is to maximize the score for negative classes. The objective function can be formulated as maximizing the following expectation of scores for negative classes:

Since LGinvolves a discrete sampling step, it cannot be directly optimized by gradientbased algorithms. Following Cai et al. [Cai and Wang (2017)], we use a one-step reinforcement learning to solve this problem.is the state,is the policy,is the action, andis the reward. We use the policy gradient to optimize the generator. From Eqs. (8) and (9), we could observe that in order to achieve higher reward, the policy used by the generator would punish the trivial negative classes by lowering down their corresponding probability. In the adversarial training, the generator and the discriminator are alternatively trained towards their respective objectives.

5 Experimental results

In this section, we first introduce the dataset and evaluation metrics, then test several variants via cross-validation to determine the parameters used in our experiments, finally show the experimental results and analysis.

5.1 Dataset and evaluation metrics

We evaluate our method on a widely used dataset4http://iesl.cs.umass.edu/riedel/ecml/that was developed by [Riedel, Yao and McCallum (2010)] and has also been used by Hoffmann et al. [Hoffmann, Zhang,Ling et al. (2011); Surdeanu, Tibshirani, Nallapati et al. (2012); Zeng, Liu, Chen et al.(2015)]. This dataset was generated by aligning Freebase relations with the NYT corpus,with sentences from the years 2005-2006 used as the training corpus and sentences from 2007 used as the testing corpus. Following the previous work [Mintz, Bills, Snow et al.(2009)], we evaluate our approach using held-out evaluation. The held-out evaluation compares the extracted relation instances against Freebase relation data.

Table 1: Parameters used in our experiments

5.2 Experimental settings

In this work, we use the Skip-gram model (word2vec)5https://code.google.com/p/word2vec/to train the word embeddings on the NYT corpus. The tokens are concatenated using the ## operator when the entity has multiple word tokens. Position features are randomly initialized with a uniform distribution between [-1, 1]. For the convenience of comparing with baseline methods,the PCNNs module uses the same parameter settings as Zeng et al. [Zeng, Liu, Chen et al.(2015)]. We use L2 regularization with regularization parameter β=0.001. We tune the proposed model using three-fold validation to study the effects of two parameters: the constant terms λ used in the loss function. We use a grid search to determine the optimal parameters and manually specify subsets of the parameter spaces:. Tab. 1 shows all parameters used in the experiments.

Figure 4: Performance comparison of proposed method and baseline methods

5.3 Baselines

We select some previous works that use handcrafted features as well as the neural network based methods as baselines. Mintz is proposed by Mintz et al. [Mintz, Bills,Snow et al. (2009)] which extracts features from all instances; MultiR is a multi-instance learning method proposed by Hoffmann et al. [Hoffmann, Zhang, Ling et al. (2011)];MIML is a multi-instance multi-labels method proposed by Surdeanu et al. [Surdeanu,Tibshirani, Nallapati et al. (2012)]; PCNNs+MIL is the method proposed by Zeng et al.[Zeng, Liu, Chen et al. (2015)], which incorporates multi-instance learning with PCNNs to extract bag features; CrossMax is proposed by Jiang et al. [Jiang, Wang, Li et al.(2016)], which exploits PCNNs and cross-sentence max-pooling to select features across different instances; Ranking Loss+Cost-Sensitive (RC) in the Zeng et al. [Zeng, Zeng and Dai (2017)], is used to address the problem of NA noise and class imbalance.

5.4 Comparison with baseline methods

In this section, we show the experimental results and comparisons with baseline methods.In the following experiments, we use Ours to refer to the proposed model that use the GAN-based framework.

The held-out evaluation provides an approximate measure of precision without requiring costly human evaluation. Half of the Freebase relations are used for testing. The relation instances discovered from the test articles are automatically compared with those in Freebase. For the convenience of comparing with baseline methods, the prediction results are sorted by the classification scores and a precision-recall curve is created for the positive classes.

Table 2: Precision, recall and F1 score of some relations

Fig. 4 shows the precision-recall curves of our approach and all the baselines. We can observe that our model outperforms all the baseline systems and improves the results significantly. It is worth emphasizing that the best of all baseline methods achieve a recall level of 38%. In contrast, our model is much better than the previous approach and enhances the recall to approximately 41%. The significant improvement can be contributed to the magic of our new pair-wise ranking loss function. The classifier uses pair-wise ranking loss which avoids explicitly learning the patterns for NA. Thus, our model will not trend to classify the samples as NA compared to softmax classifier and recalls more positive samples.

Furthermore, our model achieves a large improvement in precision especially at higher recall levels. From Fig. 4, we can see that our model achieves a precision of 43% when the recall is 41%. In contrast, when PCNNS, CrossMax and RC achieve such precision,the recalls are decreased to approximately 24%, 29% and 37%, respectively. Thus, our approach is advantageous from the point of view of precision. Also note that our model also shows advantages in the precision when the recall is very low compare with RC.This phenomenon is mainly due to the fact that when using adversarial training, the generator can consistently provide high-quality negative classes. Therefore, we can conclude that our model outperforms all the baseline systems and improves the results significantly in terms of both precision and recall.

5.4.1 Effects of adversarial training

In order to validate the effects of adversarial training, we compute the confusion matrix and analyze the detail of some relations in Tab. 2. From Tab. 2, we can see that: (1) It achieves better results in the majority of relations when using the adversarial training framework; (2) The F1 score have a significant improvement in /people/person/place_lived an-d /people/person/place_of_birth compared with Zeng et al. [Zeng, Zeng and Dai(2017)], mainly due to the fact that the generator supply high-qulity negative classes.Nonetheless, the GA-N-based framework helps to improve the performance in this case.The precision-recall curves with and without GAN are illustrated in Fig. 5, from which we can also observe that it brings better performance when using the GAN in DSRE.After removing the GAN, our model degrades to PCNNS+MIL+Traditional Ranking loss.In order to further validate the effects of our model, the PCNNS+MIL result is illustrated in Fig. 5. As we expected, our method can get better performance compared with PCNNS+MIL. The superiority of our approach indicates that incorporate the GAN framework in DSRE can effectively improve DSRE.

Figure 5: Effects of generative adversarial training

6 Conclusions

In this paper, we exploit a novel GAN-based framework for DSRE. In the traditional ranking based classifier, the majority of the generated negative class can be easily discriminated from positive class. To address the shortcoming, we design a neural network as the negative class generator to supply high-quality negative classes. The generator is learning to produce more and more discriminable negative classes, while the discriminator has to become better as well. We perform experiments on a widely used dataset and verify the adversarial learning approach. Experiment results show that our approach obtains state-of-the-art performance on the dataset.

Acknowledgments:This research work is supported by the National Natural Science Foundation of China (NO. 61772454, 6171101570, 61602059). Hunan Provincial Natural Science Foundation of China (No. 2017JJ3334), the Research Foundation of Education Bureau of Hunan Province, China (No. 16C0045), and the Open Project Program of the National Laboratory of Pattern Recognition (NLPR). Professor Jin Wang is the corresponding author.

Bengio, Y.; Ducharme, R.; Vincent, P.; Jauvin, C.(2003): A neural probabilistic language model. Journal of Machine Learning Research, vol. 3, pp. 1137-1155.

Bunescu, R. C.; Mooney, R. J.(2005): Subsequence kernels for relation extraction.Advances in Neural Information Processing Systems, pp. 171-178.

Cai, L.; Wang, W.(2017): Kbgan: Adversarial learning for knowledge graph embeddings. Computation and Language.

Collobert, R.; Weston, J.; Bottou, L.; Karlen, Michael; Kavukcuoglu, K. et al.(2011): Natural language processing (almost) from scratch. Journal of Machine Learning Research, vol. 12, pp. 2493-2537.

Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.;Xu, B.; Warde-Farley, D. et al.(2014):Generative adversarial nets. Advances in neural information processing systems, pp.2672-2680.

Gurusamy, R.; Subramaniam, V.(2017): A machine learning approach for mri brain tumor classification. Computers, Materials & Continua, vol. 53, no. 2, pp. 91-108.

Hoffmann, R.; Zhang, C.; Ling, X.;Zettlemoyer, L.; Weld, D. S.(2011): Knowledgebased weak supervision for information extraction of overlapping relations. Proceedings of the 49thAnnual Meeting of the Association for Computational Linguistics: Human Language Technologies, vol. 1, pp. 541-550.

Huang, F.; Ahuja, A.; Downey, D.; Yang, Y.; Guo, Y. et al.(2014): Learning representations for weakly supervised natural language processing tasks. Computational Linguistics, vol. 40, no. 1, pp. 85-120.

Jia, R.; Liang, P.(2017): Adversarial examples for evaluating reading comprehension systems. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 2021-2031.

Jiang, X.; Wang, Q.; Li, P.; Wang, B.(2016): Relation extraction with multi-instance multilabel convolutional neural networks. COLING 2016, 26thInternational Conference on Computational Linguistics, Proceedings of the Conference: Technical Papers, pp.1471-1480.

Kambhatla, N.(2004): Combining lexical, syntactic, and semantic features with maximum entropy models for extracting relations. Proceedings of the ACL 2004 on Interactive Poster and Demonstration Sessions, pp. 22.

Lin, Y.; Shen, S.; Liu, Z.; Luan, H.; Sun, M.(2016): Neural relation extraction with selective attention over instances. Proceedings of the 54thAnnual Meeting of the Association for Computational Linguistics, vol. 1, pp. 2124-2133.

Mikolov, T.; Chen, K.; Corrado, G.; Dean, J.(2013): Efficient estimation of word representations in vector space. Computation and Language.

Mintz, M.; Bills, S.; Snow, R.; Jurafsky, D.(2009): Distant supervision for relation extraction without labeled data. ACL’09 Proceedings of the Joint Conference of the 47thAnnual Meeting of the ACL and the 4thInternational Joint Conference on Natural Language Processing of the AFNLP, vol. 2, pp. 1003-1011.

Mirza, M.; Osindero, S.(2014): Conditional generative adversarial nets. Learning, pp. 1-7.

Nguyen, T. V. T.; Moschitti, A.(2011): End-to-end relation extraction using distant supervision from external semantic repositories. Proceedings of the 49thAnnual Meeting of the Association for Computational Linguistics: Human Language Technologies, vol. 2,pp. 277-282.

Pennington, J.; Socher, R.; Manning, C. D.(2014): Glove: Global vectors for word representation. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp. 1746-1751.

Riedel, S.; Yao, L.; McCallum, A.(2010): Modeling relations and their mentions without labeled text. Proceedings of ECML PKDD, pp. 148-163.

Santos, C. N. D.; Xiang, B.; Zhou, B.(2015): Classifying relations by ranking with convolutional neural networks. Proceedings of the 53rdAnnual Meeting of the Association for Computational Linguistics and the 7thInternational Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing,vol. 1, pp. 626-634.

Surdeanu, M.; Tibshirani, J.; Nallapati, R.; Manning, C. D.(2012): Multi-instance multi-label learning for relation extraction. Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, pp. 455-465.

Williams, R. J.(1992): Simple statistical gradient-following algorithms for connectionist reinforcement learning. Reinforcement Learning, pp. 5-32.

Yu, L.; Zhang, W.; Wang, J. Yu, Y.(2017): Seqgan: Sequence generative adversarial nets with policy gradient. Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, pp. 2852-2858.

Yuan, C.; Li, X.; Wu, Q. J.; Li, J.; Sun, X.(2017): Fingerprint liveness detection from different fingerprint materials using convolutional neural network and principal component analysis. Computers, Materials & Continua, vol. 53, no. 4, pp. 357-372.

Zelenko, D.; Aone, C.; Richardella, A.(2003): Kernel methods for relation extraction.Journal of Machine Learning Research, vol. 3, pp. 1083-1106.

Zeng, D.; Liu, K.; Chen, Y.; Zhao, J.(2015): Distant supervision for relation extraction via piecewise convolutional neural networks. Conference on Empirical Methods in Natural Language Processing, pp. 1753-1762.

Zeng, D.; Liu, K.; Lai, S.; Zhou, G.; Zhao, J.(2014): Relation classification via convolutional deep neural network. Proceedings of COLING 2014, 25thInternational Conference on Computational Linguistics: Technical Papers, pp. 2335-2344.

Zeng, D.; Zeng, J.; Dai, Y.(2017): Using cost-sensitive ranking loss to improve distant supervised relation extraction. Chinese Computational Linguistics and Natural Language Processing Based on Naturally, pp. 184-196.

Zhou, G.; Su, J.; Zhang, J.; Zhang, M.(2005): Exploring various knowledge in relation extraction. Proceedings of the 43rdAnnual Meeting on Association for Computational Linguistics, pp. 427-434.