Paper Reading AI Learner

BaRISTA: Brain Scale Informed Spatiotemporal Representation of Human Intracranial Neural Activity

2025-12-13 02:19:33
Lucine L. Oganesian, Saba Hashemi, Maryam M. Shanechi

Abstract

Intracranial recordings have opened a unique opportunity to simultaneously measure activity across multiregional networks in the human brain. Recent works have focused on developing transformer-based neurofoundation models of such recordings that can generalize across subjects and datasets. However, these recordings exhibit highly complex spatiotemporal interactions across diverse spatial scales, from the single-channel scale to the scale of brain regions. As such, there remain critical open questions regarding how best to encode spatial information and how to design self-supervision tasks that enable the learning of brain network patterns and enhance downstream decoding performance using such high-dimensional, multiregional recordings. To allow for exploring these questions, we propose a new spatiotemporal transformer model of multiregional neural activity and a corresponding self-supervised masked latent reconstruction task, designed to enable flexibility in the spatial scale used for token encoding and masking. Applying this model on publicly available multiregional intracranial electrophysiology (iEEG) data, we demonstrate that adjusting the spatial scale for both token encoding and masked reconstruction significantly impacts downstream decoding. Further, we find that spatial encoding at larger scales than channel-level encoding, which is commonly used in existing iEEG transformer models, improves downstream decoding performance. Finally, we demonstrate that our method allows for region-level token encoding while also maintaining accurate channel-level neural reconstruction. Taken together, our modeling framework enables exploration of the spatial scales used for token encoding and masking, reveals their importance towards self-supervised pretraining of neurofoundation models of multiregional human brain activity, and enhances downstream decoding performance.

Abstract (translated)

颅内记录为同时测量人类大脑多区域网络的活动提供了一个独特的机会。近期的研究集中在开发能够跨不同受试者和数据集泛化的基于变压器的基础神经模型上,这些模型处理这样的记录。然而,这种记录展示了高度复杂的时空相互作用,跨越从单通道尺度到脑区尺度的不同空间尺度。因此,关于如何最有效地编码空间信息以及如何设计自我监督任务以利用此类高维、多区域记录学习大脑网络模式并提高下游解码性能,仍存在关键的未解决的问题。 为了探索这些问题,我们提出了一种新的时空变压器模型来描述多区域神经活动,并相应地提出了一个自监督掩码潜在重构任务。该任务旨在使令牌编码和掩码的空间尺度使用具有灵活性。在公开可用的多区域颅内电生理学(iEEG)数据上应用这种模型时,我们发现调整令牌编码和掩码重建的空间尺度对下游解码有显著影响。 此外,我们发现以大于通道级别编码的大规模空间进行空间编码,这是现有iEEG变压器模型中常用的通道级编码方式,可以提高下游的解码性能。最后,我们证明了我们的方法可以在区域级进行令牌编码的同时保持准确的通道级神经重构。 总的来说,我们的建模框架使探索用于令牌编码和掩码的空间尺度成为可能,并揭示了这些空间尺度对于多区域人类大脑活动的基础神经模型的自我监督预训练的重要性,同时提高了下游解码性能。

URL

https://arxiv.org/abs/2512.12135

PDF

https://arxiv.org/pdf/2512.12135.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot