Treffer: RINDNet++: Edge Detection for Discontinuity in Reflectance, Illumination, Normal, and Depth.

Title:
RINDNet++: Edge Detection for Discontinuity in Reflectance, Illumination, Normal, and Depth.
Source:
International Journal of Computer Vision; Oct2025, Vol. 133 Issue 10, p7486-7510, 25p
Database:
Complementary Index

Weitere Informationen

As a fundamental building block in computer vision, edges, categorized by the discontinuity in surface-Reflectance, Illumination, surface-Normal or Depth, are critical for scene understanding. Despite significant progress in detecting generic or individual edge types, a holistic approach to simultaneously identifying all four edge types presents a unique challenge. In this paper, we propose RINDNet++, a novel neural network solution to address this challenge by detecting all four edge types. RINDNet++ is designed with a three-stage process that effectively leverages the distinct attributes of each edge type and their relationship. In Stage I, a common backbone extracts hierarchical features for all edges. Stage II then tailors these features for each edge type using specialized decoders. Stage III predicts the initial detection results with independent decision heads based on the enhanced features from the preceding stages. Additionally, RINDNet++ incorporates an attention module that refines edge detection by highlighting inter-type relationships, leading to enhanced edge maps. To enhance rigorous training and evaluation, we introduce the first benchmark, BSDS-RIND, incorporating annotations for all four edge types and supporting both single-scale and multi-scale testing. With the integration of state-of-the-art edge detection methods, BSDS-RIND establishes a robust framework for performance evaluation. Extensive experiments show that our proposed RINDNet++ yields promising results in comparison with the state-of-the-art approaches. Moreover, RINDNet++ excels in detecting generic edges and enhances performance in downstream applications such as shadow detection and depth estimation. [ABSTRACT FROM AUTHOR]

Copyright of International Journal of Computer Vision is the property of Springer Nature and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)