Paper Reading AI Learner

How to deal with glare for improved perception of Autonomous Vehicles

2024-04-17 02:05:05
Muhammad Z. Alam, Zeeshan Kaleem, Sousso Kelouwani

Abstract

Vision sensors are versatile and can capture a wide range of visual cues, such as color, texture, shape, and depth. This versatility, along with the relatively inexpensive availability of machine vision cameras, played an important role in adopting vision-based environment perception systems in autonomous vehicles (AVs). However, vision-based perception systems can be easily affected by glare in the presence of a bright source of light, such as the sun or the headlights of the oncoming vehicle at night or simply by light reflecting off snow or ice-covered surfaces; scenarios encountered frequently during driving. In this paper, we investigate various glare reduction techniques, including the proposed saturated pixel-aware glare reduction technique for improved performance of the computer vision (CV) tasks employed by the perception layer of AVs. We evaluate these glare reduction methods based on various performance metrics of the CV algorithms used by the perception layer. Specifically, we considered object detection, object recognition, object tracking, depth estimation, and lane detection which are crucial for autonomous driving. The experimental findings validate the efficacy of the proposed glare reduction approach, showcasing enhanced performance across diverse perception tasks and remarkable resilience against varying levels of glare.

Abstract (translated)

视觉传感器具有多功能,可以捕捉各种视觉线索,如颜色、纹理、形状和深度。这种多功能,再加上机器视觉摄像头相对较低的价格,在自动驾驶车辆(AVs)中采用基于视觉的环境感知系统发挥了重要作用。然而,基于视觉的感知系统很容易受到在明亮光源下存在的眩光的影响,例如太阳或夜间或仅仅是由于雪或冰表面反射的光线;这些情况在驾驶过程中经常遇到。在本文中,我们研究了各种眩光减除技术,包括为提高AV感知层计算机视觉(CV)任务的性能而提出的饱和像素感知眩光减除技术。我们根据CV算法使用感知层所实现的各种性能指标评估这些眩光减除方法。具体来说,我们考虑了物体检测、物体识别、物体跟踪、深度估计和车道检测,这些对于自动驾驶至关重要。实验结果证实了所提出的眩光减除方法的有效性,展示了在各种感知任务中出色的性能和对于眩光水平变化的非凡的鲁棒性。

URL

https://arxiv.org/abs/2404.10992

PDF

https://arxiv.org/pdf/2404.10992.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot