Paper Reading AI Learner

Zero-shot LLM-guided Counterfactual Generation for Text

2024-05-08 03:57:45
Amrita Bhattacharjee, Raha Moraffah, Joshua Garland, Huan Liu

Abstract

Counterfactual examples are frequently used for model development and evaluation in many natural language processing (NLP) tasks. Although methods for automated counterfactual generation have been explored, such methods depend on models such as pre-trained language models that are then fine-tuned on auxiliary, often task-specific datasets. Collecting and annotating such datasets for counterfactual generation is labor intensive and therefore, infeasible in practice. Therefore, in this work, we focus on a novel problem setting: \textit{zero-shot counterfactual generation}. To this end, we propose a structured way to utilize large language models (LLMs) as general purpose counterfactual example generators. We hypothesize that the instruction-following and textual understanding capabilities of recent LLMs can be effectively leveraged for generating high quality counterfactuals in a zero-shot manner, without requiring any training or fine-tuning. Through comprehensive experiments on various downstream tasks in natural language processing (NLP), we demonstrate the efficacy of LLMs as zero-shot counterfactual generators in evaluating and explaining black-box NLP models.

Abstract (translated)

反事实例子在许多自然语言处理(NLP)任务中常被用于模型开发和评估。尽管已经探索了自动反事实生成的方法,但这些方法依赖于诸如预训练语言模型这样的模型,然后在辅助、通常针对特定任务的辅助数据集上进行微调。因此,收集和标注这样的数据集对于反事实生成来说是不够的,在实践中是不可行的。因此,在本文中,我们关注一个新颖的问题场景:零击反事实生成。为此,我们提出了一个有结构的方法,将大型语言模型(LLMs)用作通用反事实例子生成器。我们假设,LLMs最近的指令跟随和文本理解能力可以有效地用于以零击的方式生成高质量的反事实,而无需进行训练或微调。通过对自然语言处理(NLP)中的各种下游任务的全面实验,我们证明了LLMs作为零击反事实生成器在评估和解释黑盒NLP模型方面的有效性。

URL

https://arxiv.org/abs/2405.04793

PDF

https://arxiv.org/pdf/2405.04793.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot