Abstract
Forecasting in power systems often involves multivariate time series with complex dependencies and strict privacy constraints across regions. Traditional forecasting methods require significant expert knowledge and struggle to generalize across diverse deployment scenarios. Recent advancements in pre-trained time series models offer new opportunities, but their zero-shot performance on domain-specific tasks remains limited. To address these challenges, we propose a novel MoE Encoder module that augments pretrained forecasting models by injecting a sparse mixture-of-experts layer between tokenization and encoding. This design enables two key capabilities: (1) trans forming multivariate forecasting into an expert-guided univariate task, allowing the model to effectively capture inter-variable relations, and (2) supporting localized training and lightweight parameter sharing in federated settings where raw data cannot be exchanged. Extensive experiments on public multivariate datasets demonstrate that MoE-Encoder significantly improves forecasting accuracy compared to strong baselines. We further simulate federated environments and show that transferring only MoE-Encoder parameters allows efficient adaptation to new regions, with minimal performance degradation. Our findings suggest that MoE-Encoder provides a scalable and privacy-aware extension to foundation time series models.
Abstract (translated)
电力系统中的预测通常涉及多变量时间序列,这些序列具有复杂的依赖关系,并且在不同地区之间存在严格的隐私约束。传统的预测方法需要大量的专业知识,并且难以泛化到各种部署场景中。最近,在预训练的时间序列模型方面取得了进展,这为预测任务提供了新的机会,但它们在特定领域的零样本性能仍然有限。 为了应对这些挑战,我们提出了一种新颖的MoE(混合专家)编码器模块,该模块通过在标记化和编码之间注入一个稀疏的混合专家层来增强预训练的预测模型。这种设计实现了两个关键功能:(1) 将多变量预测任务转化为由专家指导的单变量任务,从而使模型能够有效地捕捉变量之间的关系;(2) 支持联邦设置中的本地化培训和轻量级参数共享,在此情况下原始数据不能交换。 在公共多变量数据集上的广泛实验表明,与强大的基线相比,MoE-Encoder显著提高了预测准确性。我们进一步模拟了联邦环境,并展示了仅传输MoE-Encoder的参数就可以高效地适应新区域,同时保持性能下降最小化。我们的发现表明,MoE-Encoder为基础时间序列模型提供了一个可扩展且隐私意识增强的扩展方式。
URL
https://arxiv.org/abs/2601.11977