Paper Reading AI Learner

D'OH: Decoder-Only random Hypernetworks for Implicit Neural Representations

2024-03-28 06:18:12
Cameron Gordon, Lachlan Ewen MacDonald, Hemanth Saratchandran, Simon Lucey
     

Abstract

Deep implicit functions have been found to be an effective tool for efficiently encoding all manner of natural signals. Their attractiveness stems from their ability to compactly represent signals with little to no off-line training data. Instead, they leverage the implicit bias of deep networks to decouple hidden redundancies within the signal. In this paper, we explore the hypothesis that additional compression can be achieved by leveraging the redundancies that exist between layers. We propose to use a novel run-time decoder-only hypernetwork - that uses no offline training data - to better model this cross-layer parameter redundancy. Previous applications of hyper-networks with deep implicit functions have applied feed-forward encoder/decoder frameworks that rely on large offline datasets that do not generalize beyond the signals they were trained on. We instead present a strategy for the initialization of run-time deep implicit functions for single-instance signals through a Decoder-Only randomly projected Hypernetwork (D'OH). By directly changing the dimension of a latent code to approximate a target implicit neural architecture, we provide a natural way to vary the memory footprint of neural representations without the costly need for neural architecture search on a space of alternative low-rate structures.

Abstract (translated)

深度隐含函数已被证明是一种有效的工具,用于高效地编码所有类型的自然信号。它们的吸引力源于它们能够以少量的离线训练数据来压缩信号。相反,它们利用深度网络的隐含偏见来解耦信号中的隐藏冗余。在本文中,我们探讨了通过利用层之间的冗余性来获得额外压缩的可能性。我们提出了一个新型的仅运行时解码器-仅网络 - 用于更好地建模跨层参数冗余。以前使用具有深度隐含函数的过网络的应用程序依赖于大型离线训练数据集,这些数据集不适用于它们所训练的信号。我们相反提出了一个通过Decoder-Only随机投影Hypernetwork(D'OH)初始化运行时深度隐含函数的策略。通过直接将潜在代码的维度变换为近似目标隐含神经架构,我们提供了在不需要进行神经架构搜索的情况下自然地变化神经表示的内存足迹的方法。

URL

https://arxiv.org/abs/2403.19163

PDF

https://arxiv.org/pdf/2403.19163.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot