Abstract
Gait recognition, which aims at identifying individuals by their walking patterns, has recently drawn increasing research attention. However, gait recognition still suffers from the conflicts between the limited binary visual clues of the silhouette and numerous covariates with diverse scales, which brings challenges to the model's adaptiveness. In this paper, we address this conflict by developing a novel MetaGait that learns to learn an omni sample adaptive representation. Towards this goal, MetaGait injects meta-knowledge, which could guide the model to perceive sample-specific properties, into the calibration network of the attention mechanism to improve the adaptiveness from the omni-scale, omni-dimension, and omni-process perspectives. Specifically, we leverage the meta-knowledge across the entire process, where Meta Triple Attention and Meta Temporal Pooling are presented respectively to adaptively capture omni-scale dependency from spatial/channel/temporal dimensions simultaneously and to adaptively aggregate temporal information through integrating the merits of three complementary temporal aggregation methods. Extensive experiments demonstrate the state-of-the-art performance of the proposed MetaGait. On CASIA-B, we achieve rank-1 accuracy of 98.7%, 96.0%, and 89.3% under three conditions, respectively. On OU-MVLP, we achieve rank-1 accuracy of 92.4%.
Abstract (translated)
步态识别旨在通过其行走模式识别个人,但该方法仍面临着轮廓有限二进制视觉线索和众多不同尺度的协变量之间的冲突,这给模型的适应性带来了挑战。在本文中,我们旨在解决这一冲突,开发了一种新形式的 MetaGait,该模型学习学习多样本自适应表示。为实现这一目标,MetaGait 在注意力机制的校准网络中注入meta-知识,以指导模型感知样本特定的性质,以提高从多尺度、多维度和多进程的视角的适应性。具体来说,我们在整个过程中利用meta-知识,其中 Meta tripleAttention 和 Meta Temporal Pooling分别用于自适应地捕捉多尺度依赖从空间/通道/时间维度同时捕捉,并自适应地聚合时间信息,通过整合三个互补的时间聚合方法的优点。广泛实验展示了所提出的 MetaGait 的最新性能。在 CASIA-B 测试中,我们分别在不同的条件下实现了排名1的准确性98.7%、96.0%和89.3%。在 OU-MVLP 测试中,我们实现了排名1的准确性92.4%。
URL
https://arxiv.org/abs/2306.03445