Abstract
Recent works on optical flow estimation use neural networks to predict the flow field that maps positions of one image to positions of the other. These networks consist of a feature extractor, a correlation volume, and finally several refinement steps. These refinement steps mimic the iterative refinements performed by classical optimization algorithms and are usually implemented by neural layers (e.g., GRU) which are recurrently executed for a fixed and pre-determined number of steps. However, relying on a fixed number of steps may result in suboptimal performance because it is not tailored to the input data. In this paper, we introduce a novel approach for predicting the derivative of the flow using a continuous model, namely neural ordinary differential equations (ODE). One key advantage of this approach is its capacity to model an equilibrium process, dynamically adjusting the number of compute steps based on the data at hand. By following a particular neural architecture, ODE solver, and associated hyperparameters, our proposed model can replicate the exact same updates as recurrent cells used in existing works, offering greater generality. Through extensive experimental analysis on optical flow benchmarks, we demonstrate that our approach achieves an impressive improvement over baseline and existing models, all while requiring only a single refinement step.
Abstract (translated)
最近关于光流估计的工作利用神经网络来预测将一张图像的位置映射到另一张图像位置的流动场。这些网络通常包含特征提取器、相关体积(correlation volume)以及几个后续的细化步骤。这些细化步骤模仿了经典优化算法中迭代改进的过程,并且通常通过递归执行一定次数的神经层(如GRU)来实现。然而,依赖于固定数量的步骤可能导致次优性能,因为它没有针对输入数据进行定制调整。 在本文中,我们提出了一种使用连续模型——即神经常微分方程(ODE)来预测流动场导数的新方法。这种方法的一个关键优势在于它能够建模一个平衡过程,并且可以根据手头的数据动态地调整计算步骤的数量。通过遵循特定的神经网络架构、ODE求解器以及相关的超参数,我们的提议模型可以复制现有工作中使用的递归单元所执行的确切更新操作,从而提供更大的通用性。 在广泛的光流基准实验分析中,我们展示了该方法比基线和现有的模型实现了显著改进,并且只需要一个细化步骤即可实现。
URL
https://arxiv.org/abs/2506.03290