Self-Supervised Person Detection in 2D Range Data using a Calibrated
Camera
- URL: http://arxiv.org/abs/2012.08890v1
- Date: Wed, 16 Dec 2020 12:10:04 GMT
- Title: Self-Supervised Person Detection in 2D Range Data using a Calibrated
Camera
- Authors: Dan Jia and Mats Steinweg and Alexander Hermans and Bastian Leibe
- Abstract summary: We propose a method to automatically generate training labels (called pseudo-labels) for 2D LiDAR-based person detectors.
We show that self-supervised detectors, trained or fine-tuned with pseudo-labels, outperform detectors trained using manual annotations.
Our method is an effective way to improve person detectors during deployment without any additional labeling effort.
- Score: 83.31666463259849
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning is the essential building block of state-of-the-art person
detectors in 2D range data. However, only a few annotated datasets are
available for training and testing these deep networks, potentially limiting
their performance when deployed in new environments or with different LiDAR
models. We propose a method, which uses bounding boxes from an image-based
detector (e.g. Faster R-CNN) on a calibrated camera to automatically generate
training labels (called pseudo-labels) for 2D LiDAR-based person detectors.
Through experiments on the JackRabbot dataset with two detector models, DROW3
and DR-SPAAM, we show that self-supervised detectors, trained or fine-tuned
with pseudo-labels, outperform detectors trained using manual annotations from
a different dataset. Combined with robust training techniques, the
self-supervised detectors reach a performance close to the ones trained using
manual annotations. Our method is an effective way to improve person detectors
during deployment without any additional labeling effort, and we release our
source code to support relevant robotic applications.
Related papers
- VaLID: Verification as Late Integration of Detections for LiDAR-Camera Fusion [2.503388496100123]
Methods using LiDAR generally outperform those using cameras only.
We propose a model-independent late fusion method, VaLID, which validates whether each predicted bounding box is acceptable.
Our approach is model-agnostic and demonstrates state-of-the-art competitive performance even when using generic camera detectors.
arXiv Detail & Related papers (2024-09-23T20:27:10Z) - TrajSSL: Trajectory-Enhanced Semi-Supervised 3D Object Detection [59.498894868956306]
Pseudo-labeling approaches to semi-supervised learning adopt a teacher-student framework.
We leverage pre-trained motion-forecasting models to generate object trajectories on pseudo-labeled data.
Our approach improves pseudo-label quality in two distinct manners.
arXiv Detail & Related papers (2024-09-17T05:35:00Z) - Vision-Language Guidance for LiDAR-based Unsupervised 3D Object Detection [16.09503890891102]
We propose an unsupervised 3D detection approach that operates exclusively on LiDAR point clouds.
We exploit the inherent CLI-temporal knowledge of LiDAR point clouds for clustering, tracking, as well as boxtext and label refinement.
Our approach outperforms state-of-the-art unsupervised 3D object detectors on the Open dataset.
arXiv Detail & Related papers (2024-08-07T14:14:53Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - Toward unlabeled multi-view 3D pedestrian detection by generalizable AI:
techniques and performance analysis [7.414308466976969]
Generalizable AI can be used to improve multi-view 3D pedestrian detection in unlabeled target scenes.
We investigate two approaches for automatically labeling target data: pseudo-labeling using a supervised detector and automatic labeling using an untrained detector.
We show that, by using the automatic labeling approach based on an untrained detector, we can obtain superior results than directly using the untrained detector or a detector trained with an existing labeled source dataset.
arXiv Detail & Related papers (2023-08-08T18:24:53Z) - Activation to Saliency: Forming High-Quality Labels for Unsupervised
Salient Object Detection [54.92703325989853]
We propose a two-stage Activation-to-Saliency (A2S) framework that effectively generates high-quality saliency cues.
No human annotations are involved in our framework during the whole training process.
Our framework reports significant performance compared with existing USOD methods.
arXiv Detail & Related papers (2021-12-07T11:54:06Z) - Supervision by Registration and Triangulation for Landmark Detection [70.13440728689231]
We present Supervision by Registration and Triangulation (SRT), an unsupervised approach that utilizes unlabeled multi-view video to improve the accuracy and precision of landmark detectors.
Being able to utilize unlabeled data enables our detectors to learn from massive amounts of unlabeled data freely available.
arXiv Detail & Related papers (2021-01-25T02:48:21Z) - Move to See Better: Self-Improving Embodied Object Detection [35.461141354989714]
We propose a method for improving object detection in testing environments.
Our agent collects multi-view data, generates 2D and 3D pseudo-labels, and fine-tunes its detector in a self-supervised manner.
arXiv Detail & Related papers (2020-11-30T19:16:51Z) - EHSOD: CAM-Guided End-to-end Hybrid-Supervised Object Detection with
Cascade Refinement [53.69674636044927]
We present EHSOD, an end-to-end hybrid-supervised object detection system.
It can be trained in one shot on both fully and weakly-annotated data.
It achieves comparable results on multiple object detection benchmarks with only 30% fully-annotated data.
arXiv Detail & Related papers (2020-02-18T08:04:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.