Paper Reading AI Learner

Compressed Heterogeneous Graph for Abstractive Multi-Document Summarization

2023-03-12 04:23:54
Miao Li, Jianzhong Qi, Jey Han Lau

Abstract

Multi-document summarization (MDS) aims to generate a summary for a number of related documents. We propose HGSUM, an MDS model that extends an encoder-decoder architecture, to incorporate a heterogeneous graph to represent different semantic units (e.g., words and sentences) of the documents. This contrasts with existing MDS models which do not consider different edge types of graphs and as such do not capture the diversity of relationships in the documents. To preserve only key information and relationships of the documents in the heterogeneous graph, HGSUM uses graph pooling to compress the input graph. And to guide HGSUM to learn compression, we introduce an additional objective that maximizes the similarity between the compressed graph and the graph constructed from the ground-truth summary during training. HGSUM is trained end-to-end with graph similarity and standard cross-entropy objectives. Experimental results over MULTI-NEWS, WCEP-100, and ARXIV show that HGSUM outperforms state-of-the-art MDS models. The code for our model and experiments is available at: this https URL.

Abstract (translated)

多文档摘要(MDS)的目标是为多个相关文档生成摘要。我们提出了HGSUM,一个扩展编码-解码架构的MDS模型,以引入一种 heterogeneous graph 来代表文档的不同语义单元(例如,单词和句子)。这与现有的MDS模型不同,它们不考虑不同的图形边类型,因此无法捕捉文档中关系的多样性。为了在 heterogeneous graph 中保留文档的关键信息和关系,HGSUM使用图聚合来压缩输入图。为了指导HGSUM学习压缩,我们引入了另一个目标,该目标最大化训练期间压缩图与实际摘要生成图之间的相似性。HGSUM以图形相似性和标准交叉熵目标进行训练。在Multi-NEWS、WCEP-100和ARXIV的实验结果中,表明HGSUM比最先进的MDS模型表现更好。我们的模型和实验代码可在以下httpsURL获得。

URL

https://arxiv.org/abs/2303.06565

PDF

https://arxiv.org/pdf/2303.06565.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot