Paper Reading AI Learner

Drawing the Line: Deep Segmentation for Extracting Art from Ancient Etruscan Mirrors

2024-04-24 14:57:37
Rafael Sterzinger, Simon Brenner, Robert Sablatnig

Abstract

Etruscan mirrors constitute a significant category within Etruscan art and, therefore, undergo systematic examinations to obtain insights into ancient times. A crucial aspect of their analysis involves the labor-intensive task of manually tracing engravings from the backside. Additionally, this task is inherently challenging due to the damage these mirrors have sustained, introducing subjectivity into the process. We address these challenges by automating the process through photometric-stereo scanning in conjunction with deep segmentation networks which, however, requires effective usage of the limited data at hand. We accomplish this by incorporating predictions on a per-patch level, and various data augmentations, as well as exploring self-supervised learning. Compared to our baseline, we improve predictive performance w.r.t. the pseudo-F-Measure by around 16%. When assessing performance on complete mirrors against a human baseline, our approach yields quantitative similar performance to a human annotator and significantly outperforms existing binarization methods. With our proposed methodology, we streamline the annotation process, enhance its objectivity, and reduce overall workload, offering a valuable contribution to the examination of these historical artifacts and other non-traditional documents.

Abstract (translated)

伊特鲁里亚镜子在伊特鲁里亚艺术中是一个重要的分类,因此它们会进行系统性的检查,以获得对古代时代的洞察。分析的关键部分涉及从背面手动绘制浮雕的劳动密集型任务。此外,由于这些镜子所承受的损害,这项任务本质上具有挑战性,并引入了主观因素。我们通过将光栅化与深度分割网络相结合来自动化这个过程,尽管如此,需要有效使用手头有限的数据。我们通过在每片区域上预测以及各种数据增强方法以及探索自监督学习来实现这一目标。与我们的基线相比,我们在关于伪-F-测量方面的预测性能提高了约16%。当评估整体镜子与人类基线上的表现时,我们的方法与人类注释者的量化类似,显著超过了现有的二值化方法。凭借我们提出的方法,我们简化了注释过程,提高了其客观性,并降低了整体工作量,为研究这些历史文物以及其他非传统文献提供了宝贵的贡献。

URL

https://arxiv.org/abs/2404.15903

PDF

https://arxiv.org/pdf/2404.15903.pdf


Tags
3D Action Action_Localization Action_Recognition Activity Adversarial Agent Attention Autonomous Bert Boundary_Detection Caption Chat Classification CNN Compressive_Sensing Contour Contrastive_Learning Deep_Learning Denoising Detection Dialog Diffusion Drone Dynamic_Memory_Network Edge_Detection Embedding Embodied Emotion Enhancement Face Face_Detection Face_Recognition Facial_Landmark Few-Shot Gait_Recognition GAN Gaze_Estimation Gesture Gradient_Descent Handwriting Human_Parsing Image_Caption Image_Classification Image_Compression Image_Enhancement Image_Generation Image_Matting Image_Retrieval Inference Inpainting Intelligent_Chip Knowledge Knowledge_Graph Language_Model LLM Matching Medical Memory_Networks Multi_Modal Multi_Task NAS NMT Object_Detection Object_Tracking OCR Ontology Optical_Character Optical_Flow Optimization Person_Re-identification Point_Cloud Portrait_Generation Pose Pose_Estimation Prediction QA Quantitative Quantitative_Finance Quantization Re-identification Recognition Recommendation Reconstruction Regularization Reinforcement_Learning Relation Relation_Extraction Represenation Represenation_Learning Restoration Review RNN Robot Salient Scene_Classification Scene_Generation Scene_Parsing Scene_Text Segmentation Self-Supervised Semantic_Instance_Segmentation Semantic_Segmentation Semi_Global Semi_Supervised Sence_graph Sentiment Sentiment_Classification Sketch SLAM Sparse Speech Speech_Recognition Style_Transfer Summarization Super_Resolution Surveillance Survey Text_Classification Text_Generation Tracking Transfer_Learning Transformer Unsupervised Video_Caption Video_Classification Video_Indexing Video_Prediction Video_Retrieval Visual_Relation VQA Weakly_Supervised Zero-Shot