360$^\circ$ from a Single Camera: A Few-Shot Approach for LiDAR
Segmentation
- URL: http://arxiv.org/abs/2309.06197v1
- Date: Tue, 12 Sep 2023 13:04:41 GMT
- Title: 360$^\circ$ from a Single Camera: A Few-Shot Approach for LiDAR
Segmentation
- Authors: Laurenz Reichardt, Nikolas Ebert, Oliver Wasenm\"uller
- Abstract summary: Deep learning applications on LiDAR data suffer from a strong domain gap when applied to different sensors or tasks.
In practical applications labeled data is costly and time consuming to obtain.
We propose ImageTo360, an effective and streamlined few-shot approach to label-efficient LiDAR segmentation.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning applications on LiDAR data suffer from a strong domain gap when
applied to different sensors or tasks. In order for these methods to obtain
similar accuracy on different data in comparison to values reported on public
benchmarks, a large scale annotated dataset is necessary. However, in practical
applications labeled data is costly and time consuming to obtain. Such factors
have triggered various research in label-efficient methods, but a large gap
remains to their fully-supervised counterparts. Thus, we propose ImageTo360, an
effective and streamlined few-shot approach to label-efficient LiDAR
segmentation. Our method utilizes an image teacher network to generate semantic
predictions for LiDAR data within a single camera view. The teacher is used to
pretrain the LiDAR segmentation student network, prior to optional fine-tuning
on 360$^\circ$ data. Our method is implemented in a modular manner on the point
level and as such is generalizable to different architectures. We improve over
the current state-of-the-art results for label-efficient methods and even
surpass some traditional fully-supervised segmentation networks.
Related papers
- Finetuning Pre-trained Model with Limited Data for LiDAR-based 3D Object Detection by Bridging Domain Gaps [8.897884780881535]
LiDAR-based 3D object detectors often fail to adapt well to target domains with different sensor configurations.
Recent studies suggest that pre-trained backbones can be learned in a self-supervised manner with large-scale unlabeled LiDAR frames.
We propose a novel method, called Domain Adaptive Distill-Tuning (DADT), to adapt a pre-trained model with limited target data.
arXiv Detail & Related papers (2024-10-02T08:22:42Z) - Weakly Contrastive Learning via Batch Instance Discrimination and Feature Clustering for Small Sample SAR ATR [7.2932563202952725]
We propose a novel framework named Batch Instance Discrimination and Feature Clustering (BIDFC)
In this framework, embedding distance between samples should be moderate because of the high similarity between samples in the SAR images.
Experimental results on the moving and stationary target acquisition and recognition (MSTAR) database indicate a 91.25% classification accuracy of our method fine-tuned on only 3.13% training data.
arXiv Detail & Related papers (2024-08-07T08:39:33Z) - Shelf-Supervised Cross-Modal Pre-Training for 3D Object Detection [52.66283064389691]
State-of-the-art 3D object detectors are often trained on massive labeled datasets.
Recent works demonstrate that self-supervised pre-training with unlabeled data can improve detection accuracy with limited labels.
We propose a shelf-supervised approach for generating zero-shot 3D bounding boxes from paired RGB and LiDAR data.
arXiv Detail & Related papers (2024-06-14T15:21:57Z) - Terrain-Informed Self-Supervised Learning: Enhancing Building Footprint Extraction from LiDAR Data with Limited Annotations [1.3243401820948064]
Building footprint maps offer promise of precise footprint extraction without extensive post-processing.
Deep learning methods face challenges in generalization and label efficiency.
We propose terrain-aware self-supervised learning tailored to remote sensing.
arXiv Detail & Related papers (2023-11-02T12:34:23Z) - Discwise Active Learning for LiDAR Semantic Segmentation [92.92714645987083]
Active learning (AL) provides a solution that can iteratively and intelligently label a dataset.
We propose a discwise approach (DiAL) where in each iteration, we query the region a single frame covers on global coordinates, labeling all frames simultaneously.
arXiv Detail & Related papers (2023-09-23T06:08:22Z) - Hierarchical Supervision and Shuffle Data Augmentation for 3D
Semi-Supervised Object Detection [90.32180043449263]
State-of-the-art 3D object detectors are usually trained on large-scale datasets with high-quality 3D annotations.
A natural remedy is to adopt semi-supervised learning (SSL) by leveraging a limited amount of labeled samples and abundant unlabeled samples.
This paper introduces a novel approach of Hierarchical Supervision and Shuffle Data Augmentation (HSSDA), which is a simple yet effective teacher-student framework.
arXiv Detail & Related papers (2023-04-04T02:09:32Z) - Adaptive Self-Training for Object Detection [13.07105239116411]
We introduce our method Self-Training for Object Detection (ASTOD)
ASTOD determines without cost a threshold value based directly on the ground value of the score histogram.
We use different views of the unlabeled images during the pseudo-labeling step to reduce the number of missed predictions.
arXiv Detail & Related papers (2022-12-07T15:10:40Z) - LESS: Label-Efficient Semantic Segmentation for LiDAR Point Clouds [62.49198183539889]
We propose a label-efficient semantic segmentation pipeline for outdoor scenes with LiDAR point clouds.
Our method co-designs an efficient labeling process with semi/weakly supervised learning.
Our proposed method is even highly competitive compared to the fully supervised counterpart with 100% labels.
arXiv Detail & Related papers (2022-10-14T19:13:36Z) - How to distribute data across tasks for meta-learning? [59.608652082495624]
We show that the optimal number of data points per task depends on the budget, but it converges to a unique constant value for large budgets.
Our results suggest a simple and efficient procedure for data collection.
arXiv Detail & Related papers (2021-03-15T15:38:47Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.