Paper Reading AI Learner

Neural Relation Extraction via Inner-Sentence Noise Reduction and Transfer Learning

2018-08-21 02:15:05
Tianyi Liu, Xinsong Zhang, Wanhao Zhou, Weijia Jia

Abstract

Extracting relations is critical for knowledge base completion and construction in which distant supervised methods are widely used to extract relational facts automatically with the existing knowledge bases. However, the automatically constructed datasets comprise amounts of low-quality sentences containing noisy words, which is neglected by current distant supervised methods resulting in unacceptable precisions. To mitigate this problem, we propose a novel word-level distant supervised approach for relation extraction. We first build Sub-Tree Parse(STP) to remove noisy words that are irrelevant to relations. Then we construct a neural network inputting the sub-tree while applying the entity-wise attention to identify the important semantic features of relational words in each instance. To make our model more robust against noisy words, we initialize our network with a priori knowledge learned from the relevant task of entity classification by transfer learning. We conduct extensive experiments using the corpora of New York Times(NYT) and Freebase. Experiments show that our approach is effective and improves the area of Precision/Recall(PR) from 0.35 to 0.39 over the state-of-the-art work.

Abstract (translated)

提取关系对于知识库的完成和构建至关重要,其中远程监督方法被广泛用于利用现有知识库自动提取关系事实。然而,自动构建的数据集包括含有噪声单词的低质量句子的数量,当前远程监督方法忽略了这些数据,导致不可接受的精确度。为了缓解这个问题,我们提出了一种用于关系提取的新颖的单词级远程监督方法。我们首先构建子树解析(STP)以移除与关系无关的嘈杂字。然后我们构建一个输入子树的神经网络,同时应用实体方面的注意来识别每个实例中关系词的重要语义特征。为了使我们的模型更能抵抗嘈杂的单词,我们使用通过转移学习从实体分类的相关任务中学到的先验知识来初始化我们的网络。我们使用纽约时报(NYT)和Freebase的语料库进行了大量的实验。实验表明,我们的方法是有效的,并且在最先进的工作中将精确/召回(PR)的面积从0.35提高到0.39。

URL

https://arxiv.org/abs/1808.06738

PDF

https://arxiv.org/pdf/1808.06738.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot