Paper Reading AI Learner

E-CHUM: Event-based Cameras for Human Detection and Urban Monitoring

2025-12-11 19:46:17
Jack Brady, Andrew Dailey, Kristen Schang, Zo Vic Shong

Abstract

Understanding human movement and city dynamics has always been challenging. From traditional methods of manually observing the city's inhabitant, to using cameras, to now using sensors and more complex technology, the field of urban monitoring has evolved greatly. Still, there are more that can be done to unlock better practices for understanding city dynamics. This paper surveys how the landscape of urban dynamics studying has evolved with a particular focus on event-based cameras. Event-based cameras capture changes in light intensity instead of the RGB values that traditional cameras do. They offer unique abilities, like the ability to work in low-light, that can make them advantageous compared to other sensors. Through an analysis of event-based cameras, their applications, their advantages and challenges, and machine learning applications, we propose event-based cameras as a medium for capturing information to study urban dynamics. They offer the ability to capture important information while maintaining privacy. We also suggest multi-sensor fusion of event-based cameras and other sensors in the study of urban dynamics. Combining event-based cameras and infrared, event-LiDAR, or vibration has to potential to enhance the ability of event-based cameras and overcome the challenges that event-based cameras have.

Abstract (translated)

理解人类运动和城市动态一直是一项挑战。从传统的手动观察城市居民,到使用相机,再到如今采用传感器和其他复杂技术,城市管理领域已经取得了巨大的进步。然而,还有更多可以做来解锁更好的实践方法以更好地了解城市动态。本文回顾了研究城市动态领域的演变,并特别关注事件驱动型摄像头(event-based cameras)。事件驱动型摄像头捕捉的是光线强度的变化而非传统相机所捕捉的RGB值。它们具有独特的功能,例如在低光条件下工作的能力,这使它们相较于其他传感器更有优势。通过对事件驱动型摄像头及其应用、优点和挑战以及机器学习应用进行分析,我们提出将事件驱动型摄像头作为研究城市动态的信息采集手段。这种技术能够捕捉重要信息的同时维护隐私。此外,我们也建议结合使用事件驱动型摄像头和其他传感器(如红外线、事件激光雷达或振动)来研究城市动态,以增强事件驱动型摄像头的能力并克服其面临的挑战。

URL

https://arxiv.org/abs/2512.11076

PDF

https://arxiv.org/pdf/2512.11076.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Time_Series Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot