Paper Reading AI Learner

Enhancing Fingerprint Image Synthesis with GANs, Diffusion Models, and Style Transfer Techniques

2024-03-20 18:36:30
W. Tang, D. Figueroa, D. Liu, K. Johnsson, A. Sopasakis

Abstract

We present novel approaches involving generative adversarial networks and diffusion models in order to synthesize high quality, live and spoof fingerprint images while preserving features such as uniqueness and diversity. We generate live fingerprints from noise with a variety of methods, and we use image translation techniques to translate live fingerprint images to spoof. To generate different types of spoof images based on limited training data we incorporate style transfer techniques through a cycle autoencoder equipped with a Wasserstein metric along with Gradient Penalty (CycleWGAN-GP) in order to avoid mode collapse and instability. We find that when the spoof training data includes distinct spoof characteristics, it leads to improved live-to-spoof translation. We assess the diversity and realism of the generated live fingerprint images mainly through the Fréchet Inception Distance (FID) and the False Acceptance Rate (FAR). Our best diffusion model achieved a FID of 15.78. The comparable WGAN-GP model achieved slightly higher FID while performing better in the uniqueness assessment due to a slightly lower FAR when matched against the training data, indicating better creativity. Moreover, we give example images showing that a DDPM model clearly can generate realistic fingerprint images.

Abstract (translated)

我们提出了涉及生成对抗网络(GAN)和扩散模型的新颖方法,以在保留独特性和多样性特征的同时合成高质量、活体和假体指纹图像。我们使用多种方法从噪声中生成活指纹,并使用图像转换技术将活指纹图像转换为假体。为了根据有限训练数据生成不同类型的假体图像,我们在循环自动编码器(CycleAE)上配备了Wasserstein度量(Wasserstein)和梯度惩罚(CycleWGAN-GP),以避免模式坍塌和稳定性问题。我们发现,当假体训练数据包括显著的假体特征时,会导致活体到假体的翻译更好。我们通过费舍尔切比雪夫距离(FID)和假体接受率(FAR)来评估生成的活指纹图像的多样性和逼真度。我们的最佳扩散模型获得了15.78的FID。与训练数据上 slightly lower FAR 的 WGAN-GP 模型相比,具有更高的独特性评估成绩,表明在创意上表现更好。此外,我们还提供了生成真实指纹图像的示例图像,说明DDPM模型可以生成清晰逼真的指纹图像。

URL

https://arxiv.org/abs/2403.13916

PDF

https://arxiv.org/pdf/2403.13916.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot