Paper Reading AI Learner

Memorization Inheritance in Sequence-Level Knowledge Distillation for Neural Machine Translation

2025-02-03 16:26:06
Verna Dankers, Vikas Raunak

Abstract

In this work, we explore how instance-level memorization in the teacher Neural Machine Translation (NMT) model gets inherited by the student model in sequence-level knowledge distillation (SeqKD). We find that despite not directly seeing the original training data, students memorize more than baseline models (models of the same size, trained on the original data) -- 3.4% for exact matches and 57% for extractive memorization -- and show increased hallucination rates. Further, under this SeqKD setting, we also characterize how students behave on specific training data subgroups, such as subgroups with low quality and specific counterfactual memorization (CM) scores, and find that students exhibit amplified denoising on low-quality subgroups. Finally, we propose a modification to SeqKD named Adaptive-SeqKD, which intervenes in SeqKD to reduce memorization and hallucinations. Overall, we recommend caution when applying SeqKD: students inherit both their teachers' superior performance and their fault modes, thereby requiring active monitoring.

Abstract (translated)

在这项工作中,我们探讨了教师神经机器翻译(NMT)模型中的实例级记忆如何通过序列级别的知识蒸馏(SeqKD)传递给学生模型。研究发现,尽管学生模型没有直接接触原始训练数据,但它们的记忆量比基线模型(与之大小相同,在原始数据上进行训练的模型)更多——精确匹配多出3.4%,提取式记忆多出57%——并且表现出更高的幻觉率。此外,在这种SeqKD设置下,我们还描述了学生模型在特定训练数据子集上的行为表现,例如质量低下的子组和具有特定反事实记忆(CM)得分的子组,并发现学生模型对低质量子组显示出更强的去噪效果。最后,我们提出了一种名为自适应SeqKD的SeqKD改进方法,通过干预减少学生的记忆量和幻觉现象。总体而言,我们在应用SeqKD时建议谨慎行事:学生不仅继承了教师的优点,也继承了他们的缺陷模式,因此需要进行积极监控。

URL

https://arxiv.org/abs/2502.01491

PDF

https://arxiv.org/pdf/2502.01491.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot