Result: Automatic Image Annotation for Mapped Features Detection

Title:
Automatic Image Annotation for Mapped Features Detection
Contributors:
Université de Technologie de Compiègne (UTC), Heuristique et Diagnostic des Systèmes Complexes [Compiègne] (Heudiasyc), Université de Technologie de Compiègne (UTC)-Centre National de la Recherche Scientifique (CNRS), École Nationale Supérieure de Techniques Avancées (ENSTA Paris), Institut Polytechnique de Paris (IP Paris)
Source:
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2024). :9367-9373
Publisher Information:
CCSD; IEEE, 2024.
Publication Year:
2024
Collection:
collection:CNRS
collection:UNIV-COMPIEGNE
collection:HEUDIASYC
collection:HDS_SYRI
collection:IP_PARIS
collection:ALLIANCE-SU
collection:ENSTA-PARIS
collection:ENSTA
collection:DEPARTEMENT-DE-MATHEMATIQUES
Subject Geographic:
Original Identifier:
HAL: hal-04829327
Document Type:
Conference conferenceObject<br />Conference papers
Language:
English
Relation:
info:eu-repo/semantics/altIdentifier/doi/10.1109/IROS58592.2024.10801773
DOI:
10.1109/IROS58592.2024.10801773
Rights:
info:eu-repo/semantics/OpenAccess
Accession Number:
edshal.hal.04829327v1
Database:
HAL

Further Information

Detecting road features is a key enabler for autonomous driving and localization. For instance, a reliable detection of poles which are widespread in road environments can improve localization. Modern deep learning-based perception systems need a significant amount of annotated data. Automatic annotation avoids time-consuming and costly manual annotation. Because automatic methods are prone to errors, managing annotation uncertainty is crucial to ensure a proper learning process. Fusing multiple annotation sources on the same dataset can be an efficient way to reduce the errors. This not only improves the quality of annotations, but also improves the learning of perception models. In this paper, we consider the fusion of three automatic annotation methods in images: feature projection from a high accuracy vector map combined with a lidar, image segmentation and lidar segmentation. Our experimental results demonstrate the significant benefits of multi-modal automatic annotation for pole detection through a comparative evaluation on manually annotated images. Finally, the resulting multi-modal fusion is used to fine-tune an object detection model for pole base detection using unlabeled data, showing overall improvements achieved by enhancing network specialization. The dataset is publicly available.