Paper Reading AI Learner

Semantic Consistency Regularization with Large Language Models for Semi-supervised Sentiment Analysis

2025-01-29 12:03:11
Kunrong Li, Xinyu Liu, Zhen Chen

Abstract

Accurate sentiment analysis of texts is crucial for a variety of applications, such as understanding customer feedback, monitoring market trends, and detecting public sentiment. However, manually annotating large sentiment corpora for supervised learning is labor-intensive and time-consuming. Therefore, it is essential and effective to develop a semi-supervised method for the sentiment analysis task. Although some methods have been proposed for semi-supervised text classification, they rely on the intrinsic information within the unlabeled data and the learning capability of the NLP model, which lack generalization ability to the sentiment analysis scenario and may prone to overfit. Inspired by the ability of pretrained Large Language Models (LLMs) in following instructions and generating coherent text, we propose a Semantic Consistency Regularization with Large Language Models (SCR) framework for semi-supervised sentiment analysis. We introduce two prompting strategies to semantically enhance unlabeled text using LLMs. The first is Entity-based Enhancement (SCR-EE), which involves extracting entities and numerical information, and querying the LLM to reconstruct the textual information. The second is Concept-based Enhancement (SCR-CE), which directly queries the LLM with the original sentence for semantic reconstruction. Subsequently, the LLM-augmented data is utilized for a consistency loss with confidence thresholding, which preserves high-quality agreement samples to provide additional supervision signals during training. Furthermore, to fully utilize the uncertain unlabeled data samples, we propose a class re-assembling strategy inspired by the class space shrinking theorem. Experiments show our method achieves remarkable performance over prior semi-supervised methods.

Abstract (translated)

准确的情感分析对于理解客户反馈、监控市场趋势和检测公众情感等众多应用至关重要。然而,为监督学习手动标注大量情感语料库既耗时又费力。因此,开发一种适用于情感分析任务的半监督方法是必要且有效的。虽然已经提出了一些用于半监督文本分类的方法,但这些方法依赖于未标记数据中的内在信息和NLP模型的学习能力,这在情感分析场景中缺乏泛化能力,并可能容易过拟合。受预训练大规模语言模型(LLM)遵循指令并生成连贯文本的能力的启发,我们提出了一种基于大规模语言模型的语义一致性正则化(SCR)框架用于半监督情感分析。我们介绍了两种提示策略,利用LLM来增强未标记文本的语义信息:一种是基于实体的增强(SCR-EE),涉及提取实体和数值信息,并查询LLM以重构文本信息;另一种是基于概念的增强(SCR-CE),直接使用原始句子查询LLM进行语义重构。随后,利用经过LLM扩充的数据应用一致性损失并采用置信度阈值筛选高质量的一致性样本,在训练过程中提供额外的监督信号。此外,为了充分利用不确定的未标记数据样本,我们提出了一种基于类别重组策略的方法,该方法受到类空间收缩定理的启发。实验结果表明,我们的方法在以前的半监督方法中取得了显著的效果。

URL

https://arxiv.org/abs/2501.17598

PDF

https://arxiv.org/pdf/2501.17598.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot