Paper Reading AI Learner

Towards Effective Model Editing for LLM Personalization

2025-12-15 18:58:15
Baixiang Huang, Limeng Cui, Jiapeng Liu, Haoran Wang, Jiawei Xu, Zhuiyue Tan, Yutong Chen, Chen Luo, Yi Liu, Kai Shu

Abstract

Personalization is becoming indispensable for LLMs to align with individual user preferences and needs. Yet current approaches are often computationally expensive, data-intensive, susceptible to catastrophic forgetting, and prone to performance degradation in multi-turn interactions or when handling implicit queries. To address these challenges, we conceptualize personalization as a model editing task and introduce Personalization Editing, a framework that applies localized edits guided by clustered preference representations. This design enables precise preference-aligned updates while preserving overall model capabilities. In addition, existing personalization benchmarks frequently rely on persona-based dialogs between LLMs rather than user-LLM interactions, or focus primarily on stylistic imitation while neglecting information-seeking tasks that require accurate recall of user-specific preferences. We introduce User Preference Question Answering (UPQA), a short-answer QA dataset constructed from in-situ user queries with varying levels of difficulty. Unlike prior benchmarks, UPQA directly evaluates a model's ability to recall and apply specific user preferences. Across experimental settings, Personalization Editing achieves higher editing accuracy and greater computational efficiency than fine-tuning, while outperforming prompting-based baselines in multi-turn conversations and implicit preference questions settings.

Abstract (translated)

个性化对于大型语言模型(LLM)来说正变得不可或缺,以使其能够与个人用户的偏好和需求保持一致。然而,目前的方法往往计算成本高昂、数据密集型,并且容易发生灾难性遗忘,在处理多轮交互或隐式查询时性能会下降。为了解决这些问题,我们将个性化视为一个模型编辑任务,并引入了“Personalization Editing”框架,该框架通过集群偏好数字表示进行局部编辑指导。这种设计能够在保持整体模型能力的同时实现精准的偏好对齐更新。 此外,现有的个性化基准测试通常依赖于基于角色的人机对话,而不是用户与LLM之间的实际交互,或者它们主要集中在风格模仿上,而忽略了需要准确回忆出特定用户偏好的信息查找任务。我们引入了“User Preference Question Answering”(UPQA),这是一个从现场用户的查询中构建的短答案问答数据集,并且具有不同难度级别的问题。与先前的基准测试相比,UPQA直接评估模型回忆和应用具体用户偏好的能力。 在各种实验设置下,“Personalization Editing”的编辑精度高于微调方法,并且计算效率更高,在多轮对话和隐式偏好查询场景中优于基于提示的方法。

URL

https://arxiv.org/abs/2512.13676

PDF

https://arxiv.org/pdf/2512.13676.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot