Abstract
Monitoring the population and movements of endangered species is an important task to wildlife conversation. Traditional tagging methods do not scale to large populations, while applying computer vision methods to camera sensor data requires re-identification (re-ID) algorithms to obtain accurate counts and moving trajectory of wildlife. However, existing re-ID methods are largely targeted at persons and cars, which have limited pose variations and constrained capture environments. This paper tries to fill the gap by introducing a novel large-scale dataset, the Amur Tiger Re-identification in the Wild (ATRW) dataset. ATRW contains over 8,000 video clips from 92 Amur tigers, with bounding box, pose keypoint, and tiger identity annotations. In contrast to typical re-ID datasets, the tigers are captured in a diverse set of unconstrained poses and lighting conditions. We demonstrate with a set of baseline algorithms that ATRW is a challenging dataset for re-ID. Lastly, we propose a novel method for tiger re-identification, which introduces precise pose parts modeling in deep neural networks to handle large pose variation of tigers, and reaches notable performance improvement over existing re-ID methods. The dataset will be public available at https://cvwc2019.github.io/ .
Abstract (translated)
监测濒危物种的数量和活动是野生动物对话的重要任务。传统的标记方法不适用于大量人群,而将计算机视觉方法应用于摄像机传感器数据需要重新识别(RE ID)算法来获得准确的野生动物数量和运动轨迹。然而,现有的重新识别方法主要针对人和车,这些人和车具有有限的姿态变化和有限的捕获环境。本文试图通过引入一个新的大规模数据集来填补这一空白,即野生阿穆尔虎再鉴定数据集。ATRW包含来自92只阿穆尔虎的8000多个视频片段,带有边界框、姿势关键点和老虎身份注释。与典型的REID数据集相比,老虎被捕获在一组不同的无约束姿势和照明条件下。最后,我们提出了一种新的老虎再识别方法,该方法在深神经网络中引入了精确的姿态部件建模,以处理老虎的大姿态变化,并与现有的REID方法相比,取得了显著的性能改进。数据集将在https://cvwc2019.github.io/上公开。
URL
https://arxiv.org/abs/1906.05586