Paper Reading AI Learner

APrompt4EM: Augmented Prompt Tuning for Generalized Entity Matching

2024-05-08 05:38:56
Yikuan Xia, Jiazun Chen, Xinchi Li, Jun Gao

Abstract

Generalized Entity Matching (GEM), which aims at judging whether two records represented in different formats refer to the same real-world entity, is an essential task in data management. The prompt tuning paradigm for pre-trained language models (PLMs), including the recent PromptEM model, effectively addresses the challenges of low-resource GEM in practical applications, offering a robust solution when labeled data is scarce. However, existing prompt tuning models for GEM face the challenges of prompt design and information gap. This paper introduces an augmented prompt tuning framework for the challenges, which consists of two main improvements. The first is an augmented contextualized soft token-based prompt tuning method that extracts a guiding soft token benefit for the PLMs' prompt tuning, and the second is a cost-effective information augmentation strategy leveraging large language models (LLMs). Our approach performs well on the low-resource GEM challenges. Extensive experiments show promising advancements of our basic model without information augmentation over existing methods based on moderate-size PLMs (average 5.24%+), and our model with information augmentation achieves comparable performance compared with fine-tuned LLMs, using less than 14% of the API fee.

Abstract (translated)

通用实体匹配(GEM)是一种在数据管理中判断两个以不同格式表示的记录是否指向相同现实实体的重要任务。预训练语言模型(PLMs)的提示调整范式包括最近的PromptEM模型有效地解决了实际应用中低资源GEM的挑战,为标注数据有限的情况提供了一个稳健的解决方案。然而,现有的GEM提示调整模型面临提示设计和信息差距的挑战。本文介绍了一种增强提示调整框架,用于解决这些挑战,包括两个主要改进。第一个是增加基于上下文软标记的提示调整方法,为PLMs的提示调整提供指导软标记的利益;第二个是利用大型语言模型(LLMs)进行信息增强的策略。我们的方法在低资源GEM挑战中表现出色。大量的实验结果表明,与基于中等规模PLM的现有方法相比,我们的基本模型在信息增强方面的表现具有显著的进步(平均5.24%+),而我们的带有信息增强的模型与微调的LLM的性能相当,使用了不到14%的API费用。

URL

https://arxiv.org/abs/2405.04820

PDF

https://arxiv.org/pdf/2405.04820.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot