Paper Reading AI Learner

Cross-Modal Representational Knowledge Distillation for Enhanced Spike-Informed LFP Modeling

2025-12-13 21:20:13
Eray Erturk, Saba Hashemi, Maryam M. Shanechi

Abstract

Local field potentials (LFPs) can be routinely recorded alongside spiking activity in intracortical neural experiments, measure a larger complementary spatiotemporal scale of brain activity for scientific inquiry, and can offer practical advantages over spikes, including greater long-term stability, robustness to electrode degradation, and lower power requirements. Despite these advantages, recent neural modeling frameworks have largely focused on spiking activity since LFP signals pose inherent modeling challenges due to their aggregate, population-level nature, often leading to lower predictive power for downstream task variables such as motor behavior. To address this challenge, we introduce a cross-modal knowledge distillation framework that transfers high-fidelity representational knowledge from pretrained multi-session spike transformer models to LFP transformer models. Specifically, we first train a teacher spike model across multiple recording sessions using a masked autoencoding objective with a session-specific neural tokenization strategy. We then align the latent representations of the student LFP model to those of the teacher spike model. Our results show that the Distilled LFP models consistently outperform single- and multi-session LFP baselines in both fully unsupervised and supervised settings, and can generalize to other sessions without additional distillation while maintaining superior performance. These findings demonstrate that cross-modal knowledge distillation is a powerful and scalable approach for leveraging high-performing spike models to develop more accurate LFP models.

Abstract (translated)

局部场电位(LFPs)可以在皮层神经实验中与脉冲活动常规记录在一起,能够测量更大范围的空间和时间尺度的大脑活动,为科学研究提供补充。相比脉冲活动,LFP信号具有更高的长期稳定性、对电极退化的鲁棒性以及更低的功耗等实用优势。尽管有这些优点,最近的神经建模框架大多集中在脉冲活动上,因为LFP信号因其群体水平和聚合性质而给建模带来固有的挑战,这通常会导致对未来任务变量(如运动行为)预测能力较低。 为了解决这一问题,我们引入了一种跨模式知识蒸馏框架,该框架将预训练的多会话脉冲变换模型中的高保真表示知识转移到LFP变换模型中。具体来说,首先使用带有特定于每个会话的神经标记策略的掩码自动编码目标,在多个记录会话上对教师脉冲模型进行训练。然后使学生LFP模型的潜在表征与教师脉冲模型的相一致。 我们的结果显示,蒸馏后的LFP模型在完全无监督和有监督设置中都优于单个会话和多会话LFP基线,并且能够推广到其他会话,而无需进一步的蒸馏同时保持优越性能。这些发现表明跨模式知识蒸馏是一种强大且可扩展的方法,用于利用高性能脉冲模型来开发更准确的LFP模型。

URL

https://arxiv.org/abs/2512.12461

PDF

https://arxiv.org/pdf/2512.12461.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot