Paper Reading AI Learner

GraphERE: Jointly Multiple Event-Event Relation Extraction via Graph-Enhanced Event Embeddings

2024-03-19 07:50:32
Haochen Li, Di Geng

Abstract

Events describe the state changes of entities. In a document, multiple events are connected by various relations (e.g., Coreference, Temporal, Causal, and Subevent). Therefore, obtaining the connections between events through Event-Event Relation Extraction (ERE) is critical to understand natural language. There are two main problems in the current ERE works: a. Only embeddings of the event triggers are used for event feature representation, ignoring event arguments (e.g., time, place, person, etc.) and their structure within the event. b. The interconnection between relations (e.g., temporal and causal relations usually interact with each other ) is ignored. To solve the above problems, this paper proposes a jointly multiple ERE framework called GraphERE based on Graph-enhanced Event Embeddings. First, we enrich the event embeddings with event argument and structure features by using static AMR graphs and IE graphs; Then, to jointly extract multiple event relations, we use Node Transformer and construct Task-specific Dynamic Event Graphs for each type of relation. Finally, we used a multi-task learning strategy to train the whole framework. Experimental results on the latest MAVEN-ERE dataset validate that GraphERE significantly outperforms existing methods. Further analyses indicate the effectiveness of the graph-enhanced event embeddings and the joint extraction strategy.

Abstract (translated)

事件描述了实体状态的变化。在文档中,多个事件通过各种关系(如共指、时间、因果关系和子事件)相互连接。因此,通过事件-事件关系提取(ERE)获得事件之间的连接对于理解自然语言非常重要。当前ERE工作的两个主要问题是:a. 只有事件嵌入被用于事件特征表示,忽略了事件参数(例如,时间、地点、人物等)及其在事件中的结构。b. 关系之间的相互作用(例如,时间和因果关系通常相互交互)被忽视。为解决上述问题,本文提出了一种基于图增强事件嵌入的多任务框架,称为GraphERE。首先,我们通过使用静态AMR图和IE图丰富事件嵌入;然后,为了共同提取多个事件关系,我们使用节点Transformer并为每种关系构建了任务特定的动态事件图。最后,我们使用多任务学习策略训练整个框架。对最新MAVEN-ERE数据集的实验结果证实,GraphERE显著优于现有方法。进一步的分析表明,图形增强的事件嵌入和联合提取策略的有效性。

URL

https://arxiv.org/abs/2403.12523

PDF

https://arxiv.org/pdf/2403.12523.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot