Paper Reading AI Learner

DEJA VU: Continual Model Generalization For Unseen Domains

2023-01-25 05:56:53
Chenxi Liu, Lixu Wang, Lingjuan Lyu, Chen Sun, Xiao Wang, Qi Zhu

Abstract

In real-world applications, deep learning models often run in non-stationary environments where the target data distribution continually shifts over time. There have been numerous domain adaptation (DA) methods in both online and offline modes to improve cross-domain adaptation ability. However, these DA methods typically only provide good performance after a long period of adaptation, and perform poorly on new domains before and during adaptation - in what we call the "Unfamiliar Period", especially when domain shifts happen suddenly and significantly. On the other hand, domain generalization (DG) methods have been proposed to improve the model generalization ability on unadapted domains. However, existing DG works are ineffective for continually changing domains due to severe catastrophic forgetting of learned knowledge. To overcome these limitations of DA and DG in handling the Unfamiliar Period during continual domain shift, we propose RaTP, a framework that focuses on improving models' target domain generalization (TDG) capability, while also achieving effective target domain adaptation (TDA) capability right after training on certain domains and forgetting alleviation (FA) capability on past domains. RaTP includes a training-free data augmentation module to prepare data for TDG, a novel pseudo-labeling mechanism to provide reliable supervision for TDA, and a prototype contrastive alignment algorithm to align different domains for achieving TDG, TDA and FA. Extensive experiments on Digits, PACS, and DomainNet demonstrate that RaTP significantly outperforms state-of-the-art works from Continual DA, Source-Free DA, Test-Time/Online DA, Single DG, Multiple DG and Unified DA&DG in TDG, and achieves comparable TDA and FA capabilities.

Abstract (translated)

在实际应用中,深度学习模型经常运行在时间不断变化的目标数据分布的非稳定环境中。在在线和离线模式下,已经有许多域适应(DA)方法来提高跨域适应能力。然而,这些DA方法通常只有在长时间的适应之后才能提供良好的性能,并且在适应过程中对新 domains 的表现很差 - 我们称之为“不熟悉的时期”,特别是当域变化突然且显著时。另一方面,已经提出了域泛化(DG)方法,以改善未适应域的模型泛化能力。然而,现有的DG方法对于不断更改的域来说效果不佳,因为严重的灾难性遗忘了学习知识。为了克服在持续域更改期间处理不熟悉的时期 DA 和 DG 的局限性,我们提出了RaTP,一个框架,专注于改善模型的目标域泛化能力(TDG)能力,同时也在训练某些域后实现有效的目标域适应能力(TDA)能力和过去域的遗忘缓解能力(FA)能力。RaTP 包括一个无需训练的数据增强模块来准备 TDG 的数据,一个新颖的伪标签机制来提供可靠的TDA监督,以及一个原型对比对齐算法来对齐不同域以实现 TDG、TDA 和FA。对digits、PACS 和 DomainNet进行了广泛的实验,表明RaTP 在 TDG 方面显著优于现有的持续 DA、无源 DA、测试时/在线 DA、单 DG、多 DG 和统一的目标域适应与域泛化方法,并实现了与 TDA 和FA 能力相当的 TDG 和遗忘缓解能力。

URL

https://arxiv.org/abs/2301.10418

PDF

https://arxiv.org/pdf/2301.10418.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot