Paper Reading AI Learner

CADDY Underwater Stereo-Vision Dataset for Human-Robot Interaction in the Context of Diver Activities

2018-07-12 23:00:06
Arturo Gomez Chavez, Andrea Ranieri, Davide Chiarella, Enrica Zereik, Anja Babić, Andreas Birk

Abstract

In this article we present a novel underwater dataset collected from several field trials within the EU FP7 project "Cognitive autonomous diving buddy (CADDY)", where an Autonomous Underwater Vehicle (AUV) was used to interact with divers and monitor their activities. To our knowledge, this is one of the first efforts to collect a large dataset in underwater environments targeting object classification, segmentation and human pose estimation tasks. The first part of the dataset contains stereo camera recordings (~10K) of divers performing hand gestures to communicate and interact with an AUV in different environmental conditions. These gestures samples serve to test the robustness of object detection and classification algorithms against underwater image distortions i.e., color attenuation and light backscatter. The second part includes stereo footage (~12.7K) of divers free-swimming in front of the AUV, along with synchronized IMUs measurements located throughout the diver's suit (DiverNet) which serve as ground-truth for human pose and tracking methods. In both cases, these rectified images allow investigation of 3D representation and reasoning pipelines from low-texture targets commonly present in underwater scenarios. In this paper we describe our recording platform, sensor calibration procedure plus the data format and the utilities provided to use the dataset.

Abstract (translated)

在本文中,我们提出了一个新的水下数据集,该数据集来自欧盟FP7项目“认知自主潜水伙伴(CADDY)”中的几个现场试验,其中使用自主水下航行器(AUV)与潜水员互动并监控他们的活动。据我们所知,这是在水下环境中收集大型数据集的首批努力之一,目标是对象分类,分割和人体姿势估计任务。数据集的第一部分包含潜水员执行手势以在不同环境条件下与AUV进行通信和交互的立体摄像机记录(~10K)。这些手势样本用于测试物体检测和分类算法对水下图像失真的稳健性,即颜色衰减和光反向散射。第二部分包括在AUV前面潜水员自由游泳的立体声镜头(~12.7K),以及遍布潜水员服装(DiverNet)的同步IMU测量,其作为人体姿势和跟踪方法的地面实况。在这两种情况下,这些经过校正的图像可以调查来自水下情景中常见的低纹理目标的3D表示和推理管道。在本文中,我们描述了我们的记录平台,传感器校准程序以及数据格式和为使用数据集而提供的实用程序。

URL

https://arxiv.org/abs/1807.04856

PDF

https://arxiv.org/pdf/1807.04856.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot