Paper Reading AI Learner

Chasing Better Deep Image Priors between Over- and Under-parameterization

2024-10-31 17:49:44
Qiming Wu, Xiaohan Chen, Yifan Jiang, Zhangyang Wang

Abstract

Deep Neural Networks (DNNs) are well-known to act as over-parameterized deep image priors (DIP) that regularize various image inverse problems. Meanwhile, researchers also proposed extremely compact, under-parameterized image priors (e.g., deep decoder) that are strikingly competent for image restoration too, despite a loss of accuracy. These two extremes push us to think whether there exists a better solution in the middle: between over- and under-parameterized image priors, can one identify "intermediate" parameterized image priors that achieve better trade-offs between performance, efficiency, and even preserving strong transferability? Drawing inspirations from the lottery ticket hypothesis (LTH), we conjecture and study a novel "lottery image prior" (LIP) by exploiting DNN inherent sparsity, stated as: given an over-parameterized DNN-based image prior, it will contain a sparse subnetwork that can be trained in isolation, to match the original DNN's performance when being applied as a prior to various image inverse problems. Our results validate the superiority of LIPs: we can successfully locate the LIP subnetworks from over-parameterized DIPs at substantial sparsity ranges. Those LIP subnetworks significantly outperform deep decoders under comparably compact model sizes (by often fully preserving the effectiveness of their over-parameterized counterparts), and they also possess high transferability across different images as well as restoration task types. Besides, we also extend LIP to compressive sensing image reconstruction, where a pre-trained GAN generator is used as the prior (in contrast to untrained DIP or deep decoder), and confirm its validity in this setting too. To our best knowledge, this is the first time that LTH is demonstrated to be relevant in the context of inverse problems or image priors.

Abstract (translated)

深度神经网络(DNNs)作为过度参数化的深层图像先验(DIP),已知能够正则化各种图像逆问题。与此同时,研究者也提出了极其紧凑、参数不足的图像先验(例如:深解码器),尽管在准确性上有所损失,但它们对图像恢复同样表现出色。这两种极端促使我们思考是否存在一个更好的中间解决方案:在过度和参数不足的图像先验之间,能否找到实现性能、效率以及保持强迁移能力之间更好平衡的“中间”参数化图像先验?借鉴彩票假设(LTH),我们提出了并研究了一种新颖的“彩票图像先验”(LIP)概念,通过利用DNN内在稀疏性来定义:给定一个基于过度参数化的DNN的图像先验,它将包含一个可以单独训练的稀疏子网络,在应用于各种图像逆问题时可匹配原始DNN的表现。我们的结果验证了LIP的优越性:我们可以在显著的稀疏范围内成功地从过度参数化的DIP中定位到LIP子网络。这些LIP子网络在与深解码器相当紧凑的模型尺寸下表现明显更优(通常完全保留其过度参数化对应物的有效性),并且它们还具备跨不同图像和恢复任务类型的高迁移能力。此外,我们还将LIP扩展到了压缩感知图像重建中,在这里使用预训练的GAN生成器作为先验(与未经过训练的DIP或深解码器相对比),并在这种情况下也证实了其有效性。据我们所知,这是首次在逆问题或图像先验背景下证明LTH的相关性。

URL

https://arxiv.org/abs/2410.24187

PDF

https://arxiv.org/pdf/2410.24187.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot