LapSeg3D: Weakly Supervised Semantic Segmentation of Point Clouds
Representing Laparoscopic Scenes
- URL: http://arxiv.org/abs/2207.07418v1
- Date: Fri, 15 Jul 2022 11:57:14 GMT
- Title: LapSeg3D: Weakly Supervised Semantic Segmentation of Point Clouds
Representing Laparoscopic Scenes
- Authors: Benjamin Alt, Christian Kunz, Darko Katic, Rayan Younis, Rainer
J\"akel, Beat Peter M\"uller-Stich, Martin Wagner and Franziska
Mathis-Ullrich
- Abstract summary: We propose LapSeg3D, a novel approach for the voxel-wise annotation of point clouds representing surgical scenes.
As the manual annotation of training data is highly time consuming, we introduce a semi-autonomous clustering-based pipeline for the annotation of the gallbladder.
We show LapSeg3D to generalize accurately across different gallbladders and datasets recorded with different RGB-D camera systems.
- Score: 1.7941882788670036
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The semantic segmentation of surgical scenes is a prerequisite for task
automation in robot assisted interventions. We propose LapSeg3D, a novel
DNN-based approach for the voxel-wise annotation of point clouds representing
surgical scenes. As the manual annotation of training data is highly time
consuming, we introduce a semi-autonomous clustering-based pipeline for the
annotation of the gallbladder, which is used to generate segmented labels for
the DNN. When evaluated against manually annotated data, LapSeg3D achieves an
F1 score of 0.94 for gallbladder segmentation on various datasets of ex-vivo
porcine livers. We show LapSeg3D to generalize accurately across different
gallbladders and datasets recorded with different RGB-D camera systems.
Related papers
- Open-Vocabulary Octree-Graph for 3D Scene Understanding [54.11828083068082]
Octree-Graph is a novel scene representation for open-vocabulary 3D scene understanding.
An adaptive-octree structure is developed that stores semantics and depicts the occupancy of an object adjustably according to its shape.
arXiv Detail & Related papers (2024-11-25T10:14:10Z) - View-Consistent Hierarchical 3D Segmentation Using Ultrametric Feature Fields [52.08335264414515]
We learn a novel feature field within a Neural Radiance Field (NeRF) representing a 3D scene.
Our method takes view-inconsistent multi-granularity 2D segmentations as input and produces a hierarchy of 3D-consistent segmentations as output.
We evaluate our method and several baselines on synthetic datasets with multi-view images and multi-granular segmentation, showcasing improved accuracy and viewpoint-consistency.
arXiv Detail & Related papers (2024-05-30T04:14:58Z) - Learning Feature Descriptors for Pre- and Intra-operative Point Cloud
Matching for Laparoscopic Liver Registration [0.0]
In laparoscopic liver surgery (LLS), pre-operative information can be overlaid onto the intra-operative scene by registering a 3D pre-operative model to the intra-operative partial surface reconstructed from the laparoscopic video.
To assist with this task, we explore the use of learning-based feature descriptors, which, to our best knowledge, have not been explored for use in laparoscopic liver registration.
We present the LiverMatch dataset consisting of 16 preoperative models and their simulated intra-operative 3D surfaces.
We also propose the LiverMatch network designed for this task, which outputs per-point feature descriptors, visibility scores, and
arXiv Detail & Related papers (2022-11-07T16:58:39Z) - Pseudo-label Guided Cross-video Pixel Contrast for Robotic Surgical
Scene Segmentation with Limited Annotations [72.15956198507281]
We propose PGV-CL, a novel pseudo-label guided cross-video contrast learning method to boost scene segmentation.
We extensively evaluate our method on a public robotic surgery dataset EndoVis18 and a public cataract dataset CaDIS.
arXiv Detail & Related papers (2022-07-20T05:42:19Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - TraSeTR: Track-to-Segment Transformer with Contrastive Query for
Instance-level Instrument Segmentation in Robotic Surgery [60.439434751619736]
We propose TraSeTR, a Track-to-Segment Transformer that exploits tracking cues to assist surgical instrument segmentation.
TraSeTR jointly reasons about the instrument type, location, and identity with instance-level predictions.
The effectiveness of our method is demonstrated with state-of-the-art instrument type segmentation results on three public datasets.
arXiv Detail & Related papers (2022-02-17T05:52:18Z) - FUN-SIS: a Fully UNsupervised approach for Surgical Instrument
Segmentation [16.881624842773604]
We present FUN-SIS, a Fully-supervised approach for binary Surgical Instrument.
We train a per-frame segmentation model on completely unlabelled endoscopic videos, by relying on implicit motion information and instrument shape-priors.
The obtained fully-unsupervised results for surgical instrument segmentation are almost on par with the ones of fully-supervised state-of-the-art approaches.
arXiv Detail & Related papers (2022-02-16T15:32:02Z) - Temporally Constrained Neural Networks (TCNN): A framework for
semi-supervised video semantic segmentation [5.0754434714665715]
We present Temporally Constrained Neural Networks (TCNN), a semi-supervised framework used for video semantic segmentation of surgical videos.
In this work, we show that autoencoder networks can be used to efficiently provide both spatial and temporal supervisory signals.
We demonstrate that lower-dimensional representations of predicted masks can be leveraged to provide a consistent improvement on both sparsely labeled datasets.
arXiv Detail & Related papers (2021-12-27T18:06:12Z) - Interactive Segmentation for COVID-19 Infection Quantification on
Longitudinal CT scans [40.721386089781895]
Consistent segmentation of COVID-19 patient's CT scans across multiple time points is essential to assess disease progression and response to therapy accurately.
Existing automatic and interactive segmentation models for medical images only use data from a single time point (static)
We propose a new single network model for interactive segmentation that fully utilizes all available past information to refine the segmentation of follow-up scans.
arXiv Detail & Related papers (2021-10-03T08:06:38Z) - Weakly-supervised Learning For Catheter Segmentation in 3D Frustum
Ultrasound [74.22397862400177]
We propose a novel Frustum ultrasound based catheter segmentation method.
The proposed method achieved the state-of-the-art performance with an efficiency of 0.25 second per volume.
arXiv Detail & Related papers (2020-10-19T13:56:22Z) - Self-Supervised Feature Extraction for 3D Axon Segmentation [7.181047714452116]
Existing learning-based methods to automatically trace axons in 3D brain imagery often rely on manually annotated segmentation labels.
We propose a self-supervised auxiliary task that utilizes the tube-like structure of axons to build a feature extractor from unlabeled data.
We demonstrate improved segmentation performance over the 3D U-Net model on both the SHIELD PVGPe dataset and the BigNeuron Project, single neuron Janelia dataset.
arXiv Detail & Related papers (2020-04-20T20:46:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.