Correcting and Quantifying Systematic Errors in 3D Box Annotations for Autonomous Driving
- URL: http://arxiv.org/abs/2601.14038v1
- Date: Tue, 20 Jan 2026 14:57:48 GMT
- Title: Correcting and Quantifying Systematic Errors in 3D Box Annotations for Autonomous Driving
- Authors: Alexandre Justo Miro, Ludvig af Klinteberg, Bogdan Timus, Aron Asefaw, Ajinkya Khoche, Thomas Gustafsson, Sina Sharif Mansouri, Masoud Daneshtalab,
- Abstract summary: 3D box annotation based on data from active sensors is challenging in dynamic scenarios.<n>Our work is the first to discover such annotation errors in widely used, publicly available datasets.<n>Our approach increases the quality of box annotations by more than 17% in these datasets.
- Score: 34.44189129139084
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accurate ground truth annotations are critical to supervised learning and evaluating the performance of autonomous vehicle systems. These vehicles are typically equipped with active sensors, such as LiDAR, which scan the environment in predefined patterns. 3D box annotation based on data from such sensors is challenging in dynamic scenarios, where objects are observed at different timestamps, hence different positions. Without proper handling of this phenomenon, systematic errors are prone to being introduced in the box annotations. Our work is the first to discover such annotation errors in widely used, publicly available datasets. Through our novel offline estimation method, we correct the annotations so that they follow physically feasible trajectories and achieve spatial and temporal consistency with the sensor data. For the first time, we define metrics for this problem; and we evaluate our method on the Argoverse 2, MAN TruckScenes, and our proprietary datasets. Our approach increases the quality of box annotations by more than 17% in these datasets. Furthermore, we quantify the annotation errors in them and find that the original annotations are misplaced by up to 2.5 m, with highly dynamic objects being the most affected. Finally, we test the impact of the errors in benchmarking and find that the impact is larger than the improvements that state-of-the-art methods typically achieve with respect to the previous state-of-the-art methods; showing that accurate annotations are essential for correct interpretation of performance. Our code is available at https://github.com/alexandre-justo-miro/annotation-correction-3D-boxes.
Related papers
- Mistake Attribution: Fine-Grained Mistake Understanding in Egocentric Videos [11.138754178370514]
We introduce Mistake Attribution (MATT), a task for fine-grained understanding of human mistakes in egocentric video.<n>MATT attributes mistakes to the input instruction text or the attempt video.<n>We develop MisEngine, a data engine that automatically constructs attribution-rich mistake samples from existing datasets.
arXiv Detail & Related papers (2025-11-25T17:29:12Z) - Rethinking Metrics and Benchmarks of Video Anomaly Detection [58.37571339811799]
Video Anomaly Detection (VAD) aims to detect anomalies that deviate from expectation.<n>Existing VAD metrics are influenced by single annotation bias.<n>Existing benchmarks lack the capability to evaluate scene overfitting of fully/weakly-supervised algorithms.
arXiv Detail & Related papers (2025-05-25T08:09:42Z) - Automatic Image Annotation for Mapped Features Detection [6.300346102366891]
Road features are a key enabler for autonomous driving and localization.<n>Modern deep learning-based perception systems need a significant amount of annotated data.<n>In this paper, we consider the fusion of three automatic annotation methods in images.
arXiv Detail & Related papers (2024-12-11T09:06:52Z) - Towards Zero-shot 3D Anomaly Localization [58.62650061201283]
3DzAL is a novel patch-level contrastive learning framework for 3D anomaly detection and localization.<n>We show that 3DzAL outperforms the state-of-the-art anomaly detection and localization performance.
arXiv Detail & Related papers (2024-12-05T16:25:27Z) - Uncertainty Estimation for 3D Object Detection via Evidential Learning [63.61283174146648]
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
We demonstrate both the efficacy and importance of these uncertainty estimates on identifying out-of-distribution scenes, poorly localized objects, and missing (false negative) detections.
arXiv Detail & Related papers (2024-10-31T13:13:32Z) - Are Dense Labels Always Necessary for 3D Object Detection from Point
Cloud? [72.40353149833109]
Current state-of-the-art (SOTA) 3D object detection methods often require a large amount of 3D bounding box annotations for training.
We propose a novel sparsely-annotated framework, in which we just annotate one 3D object per scene.
We develop the SS3D++ method that alternatively improves 3D detector training and confident fully-annotated scene generation.
arXiv Detail & Related papers (2024-03-05T09:38:11Z) - Map-aided annotation for pole base detection [0.0]
In this paper, a 2D HD map is used to automatically annotate pole-like features in images.
In the absence of height information, the map features are represented as pole bases at the ground level.
We show how an object detector can be trained to detect a pole base.
arXiv Detail & Related papers (2024-03-04T09:23:11Z) - Once Detected, Never Lost: Surpassing Human Performance in Offline LiDAR
based 3D Object Detection [50.959453059206446]
This paper aims for high-performance offline LiDAR-based 3D object detection.
We first observe that experienced human annotators annotate objects from a track-centric perspective.
We propose a high-performance offline detector in a track-centric perspective instead of the conventional object-centric perspective.
arXiv Detail & Related papers (2023-04-24T17:59:05Z) - Weakly Supervised Monocular 3D Object Detection using Multi-View
Projection and Direction Consistency [78.76508318592552]
Monocular 3D object detection has become a mainstream approach in automatic driving for its easy application.
Most current methods still rely on 3D point cloud data for labeling the ground truths used in the training phase.
We propose a new weakly supervised monocular 3D objection detection method, which can train the model with only 2D labels marked on images.
arXiv Detail & Related papers (2023-03-15T15:14:00Z) - Effect of Annotation Errors on Drone Detection with YOLOv3 [14.519138724931446]
In this work, different types of annotation errors for object detection problem are simulated and the performance of a popular state-of-the-art object detector, YOLOv3, is examined.
Some inevitable annotation errors in CVPR-2020 Anti-UAV Challenge dataset is also examined in this manner.
arXiv Detail & Related papers (2020-04-02T15:06:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.