3DTeethSAM: Taming SAM2 for 3D Teeth Segmentation
- URL: http://arxiv.org/abs/2512.11557v1
- Date: Fri, 12 Dec 2025 13:42:06 GMT
- Title: 3DTeethSAM: Taming SAM2 for 3D Teeth Segmentation
- Authors: Zhiguo Lu, Jianwen Lou, Mingjun Ma, Hairong Jin, Youyi Zheng, Kun Zhou,
- Abstract summary: 3DTeethSAM is an adaptation of the Segment Anything Model 2 (SAM2) for 3D teeth segmentation.<n>Our method has been validated on the 3DTeethSeg benchmark, achieving an IoU of 91.90% on high-resolution 3D teeth meshes.
- Score: 26.743010197720675
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: 3D teeth segmentation, involving the localization of tooth instances and their semantic categorization in 3D dental models, is a critical yet challenging task in digital dentistry due to the complexity of real-world dentition. In this paper, we propose 3DTeethSAM, an adaptation of the Segment Anything Model 2 (SAM2) for 3D teeth segmentation. SAM2 is a pretrained foundation model for image and video segmentation, demonstrating a strong backbone in various downstream scenarios. To adapt SAM2 for 3D teeth data, we render images of 3D teeth models from predefined views, apply SAM2 for 2D segmentation, and reconstruct 3D results using 2D-3D projections. Since SAM2's performance depends on input prompts and its initial outputs often have deficiencies, and given its class-agnostic nature, we introduce three light-weight learnable modules: (1) a prompt embedding generator to derive prompt embeddings from image embeddings for accurate mask decoding, (2) a mask refiner to enhance SAM2's initial segmentation results, and (3) a mask classifier to categorize the generated masks. Additionally, we incorporate Deformable Global Attention Plugins (DGAP) into SAM2's image encoder. The DGAP enhances both the segmentation accuracy and the speed of the training process. Our method has been validated on the 3DTeethSeg benchmark, achieving an IoU of 91.90% on high-resolution 3D teeth meshes, establishing a new state-of-the-art in the field.
Related papers
- DM-CFO: A Diffusion Model for Compositional 3D Tooth Generation with Collision-Free Optimization [20.638904379060573]
We propose an approach named DM-CFO for compositional tooth generation.<n>We show that our approach significantly improves the multiview consistency and realism of the generated teeth compared with existing methods.
arXiv Detail & Related papers (2026-03-04T00:25:09Z) - SOFTooth: Semantics-Enhanced Order-Aware Fusion for Tooth Instance Segmentation [18.381890045783376]
3D tooth instance segmentation is challenging due to crowded arches, ambiguous tooth-gingiva boundaries, missing teeth, and rare yet clinically important third molars.<n>We propose SOFTooth, a semantics-enhanced 2D-3D fusion framework that leverages frozen 2D semantics without explicit 2D mask supervision.<n>On 3DTeethSeg'22, SOFTooth achieves state-of-theart overall accuracy and mean IoU, with clear gains on cases involving third molars.
arXiv Detail & Related papers (2025-12-29T12:14:41Z) - GeoSAM2: Unleashing the Power of SAM2 for 3D Part Segmentation [81.0871900167463]
We introduce GeoSAM2, a prompt-controllable framework for 3D part segmentation.<n>Given a textureless object, we render normal and point maps from predefined viewpoints.<n>We accept simple 2D prompts - clicks or boxes - to guide part selection.<n>The predicted masks are back-projected to the object and aggregated across views.
arXiv Detail & Related papers (2025-08-19T17:58:51Z) - RadSAM: Segmenting 3D radiological images with a 2D promptable model [4.9000940389224885]
We propose RadSAM, a novel method for segmenting 3D objects with a 2D model from a single prompt.<n>We introduce a benchmark to evaluate the model's ability to segment 3D objects in CT images from a single prompt.
arXiv Detail & Related papers (2025-04-29T15:00:25Z) - DCSEG: Decoupled 3D Open-Set Segmentation using Gaussian Splatting [0.0]
We present a decoupled 3D segmentation pipeline to ensure modularity and adaptability to novel 3D representations.<n>We evaluate on synthetic and real-world indoor datasets, demonstrating improved performance over comparable NeRF-based pipelines.
arXiv Detail & Related papers (2024-12-14T21:26:44Z) - EmbodiedSAM: Online Segment Any 3D Thing in Real Time [61.2321497708998]
Embodied tasks require the agent to fully understand 3D scenes simultaneously with its exploration.<n>An online, real-time, fine-grained and highly-generalized 3D perception model is desperately needed.
arXiv Detail & Related papers (2024-08-21T17:57:06Z) - MaskSAM: Towards Auto-prompt SAM with Mask Classification for Volumetric Medical Image Segmentation [17.25946659884426]
We propose MaskSAM, a mask classification prompt-free framework for medical image segmentation.<n>Our method achieves state-of-the-art performance on AMOS2022, 90.52% Dice, which improved by 2.7% compared to nnUNet.
arXiv Detail & Related papers (2024-03-21T03:28:24Z) - 3D-U-SAM Network For Few-shot Tooth Segmentation in CBCT Images [22.86724024199165]
We propose a novel 3D-U-SAM network for 3D dental image segmentation.
In order to solve the problem of using 2D pre-trained weights on 3D datasets, we adopted a convolution approximation method.
The effectiveness of the proposed method is demonstrated in ablation experiments, comparison experiments, and sample size experiments.
arXiv Detail & Related papers (2023-09-20T02:32:09Z) - Segment Anything in 3D with Radiance Fields [83.14130158502493]
This paper generalizes the Segment Anything Model (SAM) to segment 3D objects.
We refer to the proposed solution as SA3D, short for Segment Anything in 3D.
We show in experiments that SA3D adapts to various scenes and achieves 3D segmentation within seconds.
arXiv Detail & Related papers (2023-04-24T17:57:15Z) - An Implicit Parametric Morphable Dental Model [79.29420177904022]
We present the first parametric 3D morphable dental model for both teeth and gum.
It is based on a component-wise representation for each tooth and the gum, together with a learnable latent code for each of such components.
Our reconstruction quality is on par with the most advanced global implicit representations while enabling novel applications.
arXiv Detail & Related papers (2022-11-21T12:23:54Z) - Two-Stage Mesh Deep Learning for Automated Tooth Segmentation and
Landmark Localization on 3D Intraoral Scans [56.55092443401416]
emphiMeshSegNet in the first stage of TS-MDL reached an averaged Dice similarity coefficient (DSC) at 0.953pm0.076$, significantly outperforming the original MeshSegNet.
PointNet-Reg achieved a mean absolute error (MAE) of $0.623pm0.718, mm$ in distances between the prediction and ground truth for $44$ landmarks, which is superior compared with other networks for landmark detection.
arXiv Detail & Related papers (2021-09-24T13:00:26Z) - TSGCNet: Discriminative Geometric Feature Learning with Two-Stream
GraphConvolutional Network for 3D Dental Model Segmentation [141.2690520327948]
We propose a two-stream graph convolutional network (TSGCNet) to learn multi-view information from different geometric attributes.
We evaluate our proposed TSGCNet on a real-patient dataset of dental models acquired by 3D intraoral scanners.
arXiv Detail & Related papers (2020-12-26T08:02:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.