Paper Reading AI Learner

MePo: Meta Post-Refinement for Rehearsal-Free General Continual Learnin

2026-02-08 12:15:35
Guanglong Sun, Hongwei Yan, Liyuan Wang, Zhiqi Kang, Shuang Cui, Hang Su, Jun Zhu, Yi Zhong

Abstract

To cope with uncertain changes of the external world, intelligent systems must continually learn from complex, evolving environments and respond in real time. This ability, collectively known as general continual learning (GCL), encapsulates practical challenges such as online datastreams and blurry task boundaries. Although leveraging pretrained models (PTMs) has greatly advanced conventional continual learning (CL), these methods remain limited in reconciling the diverse and temporally mixed information along a single pass, resulting in sub-optimal GCL performance. Inspired by meta-plasticity and reconstructive memory in neuroscience, we introduce here an innovative approach named Meta Post-Refinement (MePo) for PTMs-based GCL. This approach constructs pseudo task sequences from pretraining data and develops a bi-level meta-learning paradigm to refine the pretrained backbone, which serves as a prolonged pretraining phase but greatly facilitates rapid adaptation of representation learning to downstream GCL tasks. MePo further initializes a meta covariance matrix as the reference geometry of pretrained representation space, enabling GCL to exploit second-order statistics for robust output alignment. MePo serves as a plug-in strategy that achieves significant performance gains across a variety of GCL benchmarks and pretrained checkpoints in a rehearsal-free manner (e.g., 15.10\%, 13.36\%, and 12.56\% on CIFAR-100, ImageNet-R, and CUB-200 under Sup-21/1K). Our source code is available at \href{this https URL}{MePo}

Abstract (translated)

为了应对外部世界不确定的变化,智能系统必须不断地从复杂且不断演变的环境中学习,并实时响应。这种能力集体称为通用持续学习(General Continual Learning, GCL),它涵盖了诸如在线数据流和模糊的任务边界等实际挑战。虽然通过预训练模型(Pretrained Models, PTMs)已大大推进了传统的持续学习(Conventional Continual Learning, CL),但这些方法在处理单次传递中多样且时间上交错的信息时仍然受限,导致GCL性能不佳。 受到神经科学中的元可塑性(meta-plasticity)和重构记忆的启发,我们在这里介绍了一种基于PTMs进行GCL的新颖方法,称为Meta Post-Refinement (MePo)。此方法从预训练数据中构建伪任务序列,并开发了一个双层元学习范式来优化预先训练好的骨干模型。这一过程相当于延长了预训练阶段,但大大促进了向下游GCL任务的快速适应性表示学习。 MePo进一步初始化一个元协方差矩阵作为预训练表示空间的参考几何形状,使GCL能够利用二次统计信息实现稳健的输出对齐。作为一个无需回放策略的方法(rehearsal-free),MePo在各种GCL基准和预训练检查点上实现了显著的性能提升(例如,在Sup-21/1K条件下,CIFAR-100提升了15.10%,ImageNet-R提升了13.36%,CUB-200提升了12.56%)。我们的源代码可在\href{this https URL}{MePo}中获取。

URL

https://arxiv.org/abs/2602.07940

PDF

https://arxiv.org/pdf/2602.07940.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot