Paper Reading AI Learner

Underwater Color Restoration Using U-Net Denoising Autoencoder

2019-05-22 07:49:45
Yousif Hashisho, Mohamad Albadawi, Tom Krause, Uwe Freiherr von Lukas

Abstract

Visual inspection of underwater structures by vehicles, e.g. remotely operated vehicles (ROVs), plays an important role in scientific, military, and commercial sectors. However, the automatic extraction of information using software tools is hindered by the characteristics of water which degrade the quality of captured videos. As a contribution for restoring the color of underwater images, Underwater Denoising Autoencoder (UDAE) model is developed using a denoising autoencoder with U-Net architecture. The proposed network takes into consideration the accuracy and the computation cost to enable real-time implementation on underwater visual tasks using end-to-end autoencoder network. Underwater vehicles perception is improved by reconstructing captured frames; hence obtaining better performance in underwater tasks. Related learning methods use generative adversarial networks (GANs) to generate color corrected underwater images, and to our knowledge this paper is the first to deal with a single autoencoder capable of producing same or better results. Moreover, image pairs are constructed for training the proposed network, where it is hard to obtain such dataset from underwater scenery. At the end, the proposed model is compared to a state-of-the-art method.

Abstract (translated)

通过车辆(如遥控车辆)对水下结构进行目视检查,在科学、军事和商业领域发挥着重要作用。然而,利用软件工具自动提取信息受到水的特性的阻碍,水的特性会降低捕获视频的质量。为了恢复水下图像的颜色,利用U-NET结构的去噪自动编码器,建立了水下去噪自动编码器(UDAE)模型。该网络考虑了端到端自动编码网络的精度和计算成本,实现了水下视觉任务的实时实现。通过重建捕获的帧,提高了水下机器人的感知能力,从而在水下任务中获得更好的性能。相关的学习方法使用生成对抗网络(gans)生成彩色水下图像,据我们所知,本文是第一个处理能够产生相同或更好结果的单个自动编码器。此外,还构造了图像对,用于训练所提出的网络,很难从水下景色中获得这样的数据集。最后,将该模型与最新的方法进行了比较。

URL

https://arxiv.org/abs/1905.09000

PDF

https://arxiv.org/pdf/1905.09000.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot