Point-Unet: A Context-aware Point-based Neural Network for Volumetric
Segmentation
- URL: http://arxiv.org/abs/2203.08964v2
- Date: Wed, 28 Feb 2024 22:09:06 GMT
- Title: Point-Unet: A Context-aware Point-based Neural Network for Volumetric
Segmentation
- Authors: Ngoc-Vuong Ho, Tan Nguyen, Gia-Han Diep, Ngan Le, Binh-Son Hua
- Abstract summary: We propose Point-Unet, a novel method that incorporates the efficiency of deep learning with 3D point clouds into volumetric segmentation.
Our key idea is to first predict the regions of interest in the volume by learning an attentional probability map.
A comprehensive benchmark on different metrics has shown that our context-aware Point-Unet robustly outperforms the SOTA voxel-based networks.
- Score: 18.81644604997336
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical image analysis using deep learning has recently been prevalent,
showing great performance for various downstream tasks including medical image
segmentation and its sibling, volumetric image segmentation. Particularly, a
typical volumetric segmentation network strongly relies on a voxel grid
representation which treats volumetric data as a stack of individual voxel
`slices', which allows learning to segment a voxel grid to be as
straightforward as extending existing image-based segmentation networks to the
3D domain. However, using a voxel grid representation requires a large memory
footprint, expensive test-time and limiting the scalability of the solutions.
In this paper, we propose Point-Unet, a novel method that incorporates the
efficiency of deep learning with 3D point clouds into volumetric segmentation.
Our key idea is to first predict the regions of interest in the volume by
learning an attentional probability map, which is then used for sampling the
volume into a sparse point cloud that is subsequently segmented using a
point-based neural network. We have conducted the experiments on the medical
volumetric segmentation task with both a small-scale dataset Pancreas and
large-scale datasets BraTS18, BraTS19, and BraTS20 challenges. A comprehensive
benchmark on different metrics has shown that our context-aware Point-Unet
robustly outperforms the SOTA voxel-based networks at both accuracies, memory
usage during training, and time consumption during testing. Our code is
available at https://github.com/VinAIResearch/Point-Unet.
Related papers
- Contextual Embedding Learning to Enhance 2D Networks for Volumetric Image Segmentation [5.995633685952995]
2D convolutional neural networks (CNNs) can hardly exploit the spatial correlation of volumetric data.
We propose a contextual embedding learning approach to facilitate 2D CNNs capturing spatial information properly.
Our approach leverages the learned embedding and the slice-wisely neighboring matching as a soft cue to guide the network.
arXiv Detail & Related papers (2024-04-02T08:17:39Z) - PointeNet: A Lightweight Framework for Effective and Efficient Point
Cloud Analysis [28.54939134635978]
PointeNet is a network designed specifically for point cloud analysis.
Our method demonstrates flexibility by seamlessly integrating with a classification/segmentation head or embedding into off-the-shelf 3D object detection networks.
Experiments on object-level datasets, including ModelNet40, ScanObjectNN, ShapeNet KITTI, and the scene-level dataset KITTI, demonstrate the superior performance of PointeNet over state-of-the-art methods in point cloud analysis.
arXiv Detail & Related papers (2023-12-20T03:34:48Z) - Clustering based Point Cloud Representation Learning for 3D Analysis [80.88995099442374]
We propose a clustering based supervised learning scheme for point cloud analysis.
Unlike current de-facto, scene-wise training paradigm, our algorithm conducts within-class clustering on the point embedding space.
Our algorithm shows notable improvements on famous point cloud segmentation datasets.
arXiv Detail & Related papers (2023-07-27T03:42:12Z) - PointResNet: Residual Network for 3D Point Cloud Segmentation and
Classification [18.466814193413487]
Point cloud segmentation and classification are some of the primary tasks in 3D computer vision.
In this paper, we propose PointResNet, a residual block-based approach.
Our model directly processes the 3D points, using a deep neural network for the segmentation and classification tasks.
arXiv Detail & Related papers (2022-11-20T17:39:48Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - SASA: Semantics-Augmented Set Abstraction for Point-based 3D Object
Detection [78.90102636266276]
We propose a novel set abstraction method named Semantics-Augmented Set Abstraction (SASA)
Based on the estimated point-wise foreground scores, we then propose a semantics-guided point sampling algorithm to help retain more important foreground points during down-sampling.
In practice, SASA shows to be effective in identifying valuable points related to foreground objects and improving feature learning for point-based 3D detection.
arXiv Detail & Related papers (2022-01-06T08:54:47Z) - Learning point embedding for 3D data processing [2.12121796606941]
Current point-based methods are essentially spatial relationship processing networks.
Our architecture, PE-Net, learns the representation of point clouds in high-dimensional space.
Experiments show that PE-Net achieves the state-of-the-art performance in multiple challenging datasets.
arXiv Detail & Related papers (2021-07-19T00:25:28Z) - Learning Semantic Segmentation of Large-Scale Point Clouds with Random
Sampling [52.464516118826765]
We introduce RandLA-Net, an efficient and lightweight neural architecture to infer per-point semantics for large-scale point clouds.
The key to our approach is to use random point sampling instead of more complex point selection approaches.
Our RandLA-Net can process 1 million points in a single pass up to 200x faster than existing approaches.
arXiv Detail & Related papers (2021-07-06T05:08:34Z) - Compositional Prototype Network with Multi-view Comparision for Few-Shot
Point Cloud Semantic Segmentation [47.0611707526858]
A fully supervised point cloud segmentation network often requires a large amount of data with point-wise annotations.
We present the Compositional Prototype Network that can undertake point cloud segmentation with only a few labeled training data.
Inspired by the few-shot learning literature in images, our network directly transfers label information from the limited training data to unlabeled test data for prediction.
arXiv Detail & Related papers (2020-12-28T15:01:34Z) - Improving Point Cloud Semantic Segmentation by Learning 3D Object
Detection [102.62963605429508]
Point cloud semantic segmentation plays an essential role in autonomous driving.
Current 3D semantic segmentation networks focus on convolutional architectures that perform great for well represented classes.
We propose a novel Aware 3D Semantic Detection (DASS) framework that explicitly leverages localization features from an auxiliary 3D object detection task.
arXiv Detail & Related papers (2020-09-22T14:17:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.