Paper Reading AI Learner

FedStyle: Style-Based Federated Learning Crowdsourcing Framework for Art Commissions

2024-04-25 04:53:43
Changjuan Ran, Yeting Guo, Fang Liu, Shenglan Cui, Yunfan Ye

Abstract

The unique artistic style is crucial to artists' occupational competitiveness, yet prevailing Art Commission Platforms rarely support style-based retrieval. Meanwhile, the fast-growing generative AI techniques aggravate artists' concerns about releasing personal artworks to public platforms. To achieve artistic style-based retrieval without exposing personal artworks, we propose FedStyle, a style-based federated learning crowdsourcing framework. It allows artists to train local style models and share model parameters rather than artworks for collaboration. However, most artists possess a unique artistic style, resulting in severe model drift among them. FedStyle addresses such extreme data heterogeneity by having artists learn their abstract style representations and align with the server, rather than merely aggregating model parameters lacking semantics. Besides, we introduce contrastive learning to meticulously construct the style representation space, pulling artworks with similar styles closer and keeping different ones apart in the embedding space. Extensive experiments on the proposed datasets demonstrate the superiority of FedStyle.

Abstract (translated)

独特的美学风格对艺术家职业竞争力至关重要,然而现行的艺术委员会平台 rarely 支持基于风格的音乐检索。与此同时,快速增长的生成式 AI 技术使艺术家对将个人作品发布到公共平台感到担忧。为了实现基于美学风格的音乐检索而不会泄露个人作品,我们提出了 FedStyle,一种基于风格的分众学习框架。它允许艺术家训练本地风格模型并共享模型参数,而不是为了合作而共享作品。然而,大多数艺术家具有独特的艺术风格,导致他们之间的模型漂移严重。FedStyle 通过让艺术家学习其抽象风格表示来解决这种极端的数据异质性,而不是简单地聚合缺乏语义的数据参数。此外,我们还引入了对比学习来精心构建风格表示空间,将具有相似风格的作品推向更靠近,将不同风格的作品保持在空间中。在提出的数据集上进行的大量实验证明 FedStyle 的优越性。

URL

https://arxiv.org/abs/2404.16336

PDF

https://arxiv.org/pdf/2404.16336.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot