Paper Reading AI Learner

IFNet: Deep Imaging and Focusing for Handheld SAR with Millimeter-wave Signals

2024-05-03 11:55:45
Li Yadong, Zhang Dongheng, Geng Ruixu, Wu Jincheng, Hu Yang, Sun Qibin, Chen Yan

Abstract

Recent advancements have showcased the potential of handheld millimeter-wave (mmWave) imaging, which applies synthetic aperture radar (SAR) principles in portable settings. However, existing studies addressing handheld motion errors either rely on costly tracking devices or employ simplified imaging models, leading to impractical deployment or limited performance. In this paper, we present IFNet, a novel deep unfolding network that combines the strengths of signal processing models and deep neural networks to achieve robust imaging and focusing for handheld mmWave systems. We first formulate the handheld imaging model by integrating multiple priors about mmWave images and handheld phase errors. Furthermore, we transform the optimization processes into an iterative network structure for improved and efficient imaging performance. Extensive experiments demonstrate that IFNet effectively compensates for handheld phase errors and recovers high-fidelity images from severely distorted signals. In comparison with existing methods, IFNet can achieve at least 11.89 dB improvement in average peak signal-to-noise ratio (PSNR) and 64.91% improvement in average structural similarity index measure (SSIM) on a real-world dataset.

Abstract (translated)

近年来,便携式毫米波成像(mmWave Imaging)的潜在应用已经得到了展示,这种应用利用了便携式设置下的合成孔径雷达(SAR)原理。然而,现有的研究要么依赖于昂贵的跟踪设备,要么采用简化的成像模型,导致实际部署不实用或性能有限。在本文中,我们提出了IFNet,一种新颖的深度展开网络,结合了信号处理模型的优势和深度神经网络的优点,为手持mmWave系统实现稳健的成像和聚焦。我们首先通过整合多个关于mmWave图像和手持相位误差的多项prior,形式化地定义了手持成像模型。此外,我们将优化过程转化为一个迭代网络结构,以提高和实现高效的成像性能。大量实验证明IFNet能够有效补偿手持相位误差,并从严重扭曲的信号中恢复高保真的图像。与现有方法相比,IFNet可以在真实世界数据集上实现至少11.89 dB的平均峰值信号-噪声比(PSNR)的改进和64.91%的平均结构相似性指数测量(SSIM)。

URL

https://arxiv.org/abs/2405.02023

PDF

https://arxiv.org/pdf/2405.02023.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot