Paper Reading AI Learner

Propagate And Calibrate: Real-time Passive Non-line-of-sight Tracking

2023-03-21 12:18:57
Yihao Wang, Zhigang Wang, Bin Zhao, Dong Wang, Mulin Chen, Xuelong Li

Abstract

Non-line-of-sight (NLOS) tracking has drawn increasing attention in recent years, due to its ability to detect object motion out of sight. Most previous works on NLOS tracking rely on active illumination, e.g., laser, and suffer from high cost and elaborate experimental conditions. Besides, these techniques are still far from practical application due to oversimplified settings. In contrast, we propose a purely passive method to track a person walking in an invisible room by only observing a relay wall, which is more in line with real application scenarios, e.g., security. To excavate imperceptible changes in videos of the relay wall, we introduce difference frames as an essential carrier of temporal-local motion messages. In addition, we propose PAC-Net, which consists of alternating propagation and calibration, making it capable of leveraging both dynamic and static messages on a frame-level granularity. To evaluate the proposed method, we build and publish the first dynamic passive NLOS tracking dataset, NLOS-Track, which fills the vacuum of realistic NLOS datasets. NLOS-Track contains thousands of NLOS video clips and corresponding trajectories. Both real-shot and synthetic data are included.

Abstract (translated)

非可见性跟踪(NLOS)近年来吸引了越来越多的关注,因为它能够检测非可见的物体运动。以前的NLOS跟踪研究大多数依赖于 Active 照明,例如激光,并且成本高昂、实验条件复杂。此外,这些技术仍然远距应用于实际场景,例如安全。相反,我们提出了一种纯粹的 passive 方法,通过仅观察传递墙来跟踪一个人在一个看不见的房间里漫步,这更加符合实际应用场景,例如安全。为了挖掘传递墙视频中的微变化,我们引入了差异帧作为时间局部运动消息的重要载波。此外,我们提出了 PAC-Net,它包括交替传播和校准,使其能够在帧级别的 granularity 上利用动态和静态消息。为了评估提出的方法,我们建立了并发表了第一个动态 passive NLOS 跟踪数据集 NLOS-Track,该数据集填补了真实的NLOS数据集的空缺。NLOS-Track包含数千个NLOS视频片段和相应的轨迹。既有真实拍摄数据,也有合成数据。

URL

https://arxiv.org/abs/2303.11791

PDF

https://arxiv.org/pdf/2303.11791.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot