Abstract
Edge detection is a fundamental technique in various computer vision tasks. Edges are indeed effectively delineated by pixel discontinuity and can offer reliable structural information even in textureless areas. State-of-the-art heavily relies on pixel-wise annotations, which are labor-intensive and subject to inconsistencies when acquired manually. In this work, we propose a novel self-supervised approach for edge detection that employs a multi-level, multi-homography technique to transfer annotations from synthetic to real-world datasets. To fully leverage the generated edge annotations, we developed SuperEdge, a streamlined yet efficient model capable of concurrently extracting edges at pixel-level and object-level granularity. Thanks to self-supervised training, our method eliminates the dependency on manual annotated edge labels, thereby enhancing its generalizability across diverse datasets. Comparative evaluations reveal that SuperEdge advances edge detection, demonstrating improvements of 4.9% in ODS and 3.3% in OIS over the existing STEdge method on BIPEDv2.
Abstract (translated)
边缘检测是各种计算机视觉任务中的基本技术。事实上,通过像素不连续性有效地描绘边缘可以提供与纹理无关区域中的可靠结构信息。先进的边缘检测方法主要依赖于像素级的注释,这些注释工作量大且容易出现不一致。在本文中,我们提出了一种新颖的自监督边缘检测方法,该方法采用多级多同构技术将注释从合成数据集转移至现实世界数据集。为了充分利用生成的边缘注释,我们开发了SuperEdge,一种在像素级和物体级别上同时提取边缘的流线型高效模型。由于自监督训练,我们的方法消除了对手动注释边缘标签的依赖,从而提高了其对不同数据集的泛化能力。比较评估显示,SuperEdge在边缘检测方面取得了改进,其ODS提高了4.9%,OIS提高了3.3%,超过了现有STEdge方法。
URL
https://arxiv.org/abs/2401.02313