Paper Reading AI Learner

Conditional Neural Style Transfer with Peer-Regularized Feature Transform

2019-06-10 07:22:31
Jan Svoboda, Asha Anoosheh, Christian Osendorfer, Jonathan Masci

Abstract

This paper introduces a neural style transfer model to conditionally generate a stylized image using only a set of examples describing the desired style. The proposed solution produces high-quality images even in the zero-shot setting and allows for greater freedom in changing the content geometry. This is thanks to the introduction of a novel Peer-Regularization Layer that recomposes style in latent space by means of a custom graph convolutional layer aiming at separating style and content. Contrary to the vast majority of existing solutions our model does not require any pre-trained network for computing perceptual losses and can be trained fully end-to-end with a new set of cyclic losses that operate directly in latent space. An extensive ablation study confirms the usefulness of the proposed losses and of the Peer-Regularization Layer, with qualitative results that are competitive with respect to the current state-of-the-art even in the challenging zero-shot setting. This opens the door to more abstract and artistic neural image generation scenarios and easier deployment of the model in. production

Abstract (translated)

本文介绍了一种神经风格转换模型,仅通过一组描述所需风格的例子,有条件地生成一个风格化的图像。所提出的解决方案即使在零镜头设置下也能生成高质量的图像,并允许更自由地更改内容几何图形。这是由于引入了一种新的对等正则化层,该层通过一个定制的图形卷积层来重新编译潜在空间中的样式,目的是分离样式和内容。与现有的绝大多数解决方案相反,我们的模型不需要任何经过预先培训的网络来计算感知损失,并且可以通过直接在潜在空间中运行的一组新的循环损失进行端到端的全面培训。一项广泛的烧蚀研究证实了所提议的损失和对等正则化层的有用性,定性结果与当前最先进的技术相竞争,即使在具有挑战性的零触发设置中也是如此。这为更抽象和艺术的神经图像生成场景和更容易在中部署模型打开了大门。生产

URL

https://arxiv.org/abs/1906.02913

PDF

https://arxiv.org/pdf/1906.02913.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot