Interactive Object Segmentation in 3D Point Clouds
- URL: http://arxiv.org/abs/2204.07183v1
- Date: Thu, 14 Apr 2022 18:31:59 GMT
- Title: Interactive Object Segmentation in 3D Point Clouds
- Authors: Theodora Kontogianni, Ekin Celikkan, Siyu Tang and Konrad Schindler
- Abstract summary: We present an interactive 3D object segmentation method in which the user interacts directly with the 3D point cloud.
Our model does not require training data from the target domain.
It performs well on several other datasets with different data characteristics as well as different object classes.
- Score: 27.88495480980352
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning depends on large amounts of labeled training data. Manual
labeling is expensive and represents a bottleneck, especially for tasks such as
segmentation, where labels must be assigned down to the level of individual
points. That challenge is even more daunting for 3D data: 3D point clouds
contain millions of points per scene, and their accurate annotation is markedly
more time-consuming. The situation is further aggravated by the added
complexity of user interfaces for 3D point clouds, which slows down annotation
even more. For the case of 2D image segmentation, interactive techniques have
become common, where user feedback in the form of a few clicks guides a
segmentation algorithm -- nowadays usually a neural network -- to achieve an
accurate labeling with minimal effort. Surprisingly, interactive segmentation
of 3D scenes has not been explored much. Previous work has attempted to obtain
accurate 3D segmentation masks using human feedback from the 2D domain, which
is only possible if correctly aligned images are available together with the 3D
point cloud, and it involves switching between the 2D and 3D domains. Here, we
present an interactive 3D object segmentation method in which the user
interacts directly with the 3D point cloud. Importantly, our model does not
require training data from the target domain: when trained on ScanNet, it
performs well on several other datasets with different data characteristics as
well as different object classes. Moreover, our method is orthogonal to
supervised (instance) segmentation methods and can be combined with them to
refine automatic segmentations with minimal human effort.
Related papers
- Segment3D: Learning Fine-Grained Class-Agnostic 3D Segmentation without
Manual Labels [141.23836433191624]
Current 3D scene segmentation methods are heavily dependent on manually annotated 3D training datasets.
We propose Segment3D, a method for class-agnostic 3D scene segmentation that produces high-quality 3D segmentation masks.
arXiv Detail & Related papers (2023-12-28T18:57:11Z) - SAM-guided Graph Cut for 3D Instance Segmentation [60.75119991853605]
This paper addresses the challenge of 3D instance segmentation by simultaneously leveraging 3D geometric and multi-view image information.
We introduce a novel 3D-to-2D query framework to effectively exploit 2D segmentation models for 3D instance segmentation.
Our method achieves robust segmentation performance and can generalize across different types of scenes.
arXiv Detail & Related papers (2023-12-13T18:59:58Z) - DatasetNeRF: Efficient 3D-aware Data Factory with Generative Radiance Fields [68.94868475824575]
This paper introduces a novel approach capable of generating infinite, high-quality 3D-consistent 2D annotations alongside 3D point cloud segmentations.
We leverage the strong semantic prior within a 3D generative model to train a semantic decoder.
Once trained, the decoder efficiently generalizes across the latent space, enabling the generation of infinite data.
arXiv Detail & Related papers (2023-11-18T21:58:28Z) - PartSLIP: Low-Shot Part Segmentation for 3D Point Clouds via Pretrained
Image-Language Models [56.324516906160234]
Generalizable 3D part segmentation is important but challenging in vision and robotics.
This paper explores an alternative way for low-shot part segmentation of 3D point clouds by leveraging a pretrained image-language model, GLIP.
We transfer the rich knowledge from 2D to 3D through GLIP-based part detection on point cloud rendering and a novel 2D-to-3D label lifting algorithm.
arXiv Detail & Related papers (2022-12-03T06:59:01Z) - MvDeCor: Multi-view Dense Correspondence Learning for Fine-grained 3D
Segmentation [91.6658845016214]
We propose to utilize self-supervised techniques in the 2D domain for fine-grained 3D shape segmentation tasks.
We render a 3D shape from multiple views, and set up a dense correspondence learning task within the contrastive learning framework.
As a result, the learned 2D representations are view-invariant and geometrically consistent.
arXiv Detail & Related papers (2022-08-18T00:48:15Z) - Semi-supervised 3D shape segmentation with multilevel consistency and
part substitution [21.075426681857024]
We propose an effective semi-supervised method for learning 3D segmentations from a few labeled 3D shapes and a large amount of unlabeled 3D data.
For the unlabeled data, we present a novel multilevel consistency loss to enforce consistency of network predictions between perturbed copies of a 3D shape.
For the labeled data, we develop a simple yet effective part substitution scheme to augment the labeled 3D shapes with more structural variations to enhance training.
arXiv Detail & Related papers (2022-04-19T11:48:24Z) - Learning 3D Semantic Segmentation with only 2D Image Supervision [18.785840615548473]
We train a 3D model from pseudo-labels derived from 2D semantic image segmentations using multiview fusion.
The proposed network architecture, 2D3DNet, achieves significantly better performance than baselines during experiments on a new urban dataset with lidar and images captured in 20 cities across 5 continents.
arXiv Detail & Related papers (2021-10-21T17:56:28Z) - Interactive Annotation of 3D Object Geometry using 2D Scribbles [84.51514043814066]
In this paper, we propose an interactive framework for annotating 3D object geometry from point cloud data and RGB imagery.
Our framework targets naive users without artistic or graphics expertise.
arXiv Detail & Related papers (2020-08-24T21:51:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.