Paper Reading AI Learner

Unconditional stability of a recurrent neural circuit implementing divisive normalization

2024-09-27 17:46:05
Shivang Rawat, David J. Heeger, Stefano Martiniani

Abstract

Stability in recurrent neural models poses a significant challenge, particularly in developing biologically plausible neurodynamical models that can be seamlessly trained. Traditional cortical circuit models are notoriously difficult to train due to expansive nonlinearities in the dynamical system, leading to an optimization problem with nonlinear stability constraints that are difficult to impose. Conversely, recurrent neural networks (RNNs) excel in tasks involving sequential data but lack biological plausibility and interpretability. In this work, we address these challenges by linking dynamic divisive normalization (DN) to the stability of ORGaNICs, a biologically plausible recurrent cortical circuit model that dynamically achieves DN and has been shown to simulate a wide range of neurophysiological phenomena. By using the indirect method of Lyapunov, we prove the remarkable property of unconditional local stability for an arbitrary-dimensional ORGaNICs circuit when the recurrent weight matrix is the identity. We thus connect ORGaNICs to a system of coupled damped harmonic oscillators, which enables us to derive the circuit's energy function, providing a normative principle of what the circuit, and individual neurons, aim to accomplish. Further, for a generic recurrent weight matrix, we prove the stability of the 2D model and demonstrate empirically that stability holds in higher dimensions. Finally, we show that ORGaNICs can be trained by backpropagation through time without gradient clipping/scaling, thanks to its intrinsic stability property and adaptive time constants, which address the problems of exploding, vanishing, and oscillating gradients. By evaluating the model's performance on RNN benchmarks, we find that ORGaNICs outperform alternative neurodynamical models on static image classification tasks and perform comparably to LSTMs on sequential tasks.

Abstract (translated)

循环神经网络(RNNs)的稳定性 poses 是一个重大的挑战,尤其是在开发生物 plausible 的神经动力学模型时,可以无缝训练。传统的皮质电路模型由于动态系统中的非线性拓扑结构,导致具有非线性稳定性约束的优化问题很难求解。相反,循环神经网络(RNNs)在涉及序列数据的任务上表现出色,但缺乏生物合理性和可解释性。在这项工作中,我们通过将动态分枝 normalization(DN)与 ORGaNICs 的稳定性联系起来,解决了这些挑战。ORGaNICs 是一种生物 plausible 的循环神经网络模型,具有动态地实现 DN 并已展示出模拟广泛的神经生理现象的能力。通过使用 Lyapunov 间接法,我们证明了任意维度的 ORGaNICs 电路在循环权重矩阵为 identity 时的条件局部稳定性惊人的特性。因此,我们将 ORGaNICs 连接到耦合阻尼谐波振荡器的系统中,这使得我们能够求出电路的能量函数,为电路和单个神经元提供了规范原理,即它们致力于实现的目标。此外,对于任意维度的通用循环权重矩阵,我们证明了 2D 模型的稳定性,并实验证明了在更高维度中稳定性依然存在。最后,我们证明了 ORGaNICs 可以通过反向传播通过时间进行训练,得益于其固有稳定性特性和自适应时间常数,这解决了爆炸、消失和振荡梯度的問題。通过评估模型在 RNN 基准测试上的性能,我们发现 ORGaNICs 在静态图像分类任务上优于其他神经动力学模型,同时在序列任务上与 LSTMs 相当。

URL

https://arxiv.org/abs/2409.18946

PDF

https://arxiv.org/pdf/2409.18946.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot