A Deep Learning Approach to Teeth Segmentation and Orientation from
Panoramic X-rays
- URL: http://arxiv.org/abs/2310.17176v1
- Date: Thu, 26 Oct 2023 06:01:25 GMT
- Title: A Deep Learning Approach to Teeth Segmentation and Orientation from
Panoramic X-rays
- Authors: Mrinal Kanti Dhar, Mou Deb, D. Madhab, and Zeyun Yu
- Abstract summary: We present a comprehensive approach to teeth segmentation and orientation from panoramic X-ray images, leveraging deep learning techniques.
We build our model based on FUSegNet, a popular model originally developed for wound segmentation.
We introduce oriented bounding box (OBB) generation through principal component analysis (PCA) for precise tooth orientation estimation.
- Score: 1.7366868394060984
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Accurate teeth segmentation and orientation are fundamental in modern oral
healthcare, enabling precise diagnosis, treatment planning, and dental implant
design. In this study, we present a comprehensive approach to teeth
segmentation and orientation from panoramic X-ray images, leveraging deep
learning techniques. We build our model based on FUSegNet, a popular model
originally developed for wound segmentation, and introduce modifications by
incorporating grid-based attention gates into the skip connections. We
introduce oriented bounding box (OBB) generation through principal component
analysis (PCA) for precise tooth orientation estimation. Evaluating our
approach on the publicly available DNS dataset, comprising 543 panoramic X-ray
images, we achieve the highest Intersection-over-Union (IoU) score of 82.43%
and Dice Similarity Coefficient (DSC) score of 90.37% among compared models in
teeth instance segmentation. In OBB analysis, we obtain the Rotated IoU (RIoU)
score of 82.82%. We also conduct detailed analyses of individual tooth labels
and categorical performance, shedding light on strengths and weaknesses. The
proposed model's accuracy and versatility offer promising prospects for
improving dental diagnoses, treatment planning, and personalized healthcare in
the oral domain. Our generated OBB coordinates and codes are available at
https://github.com/mrinal054/Instance_teeth_segmentation.
Related papers
- Boundary feature fusion network for tooth image segmentation [7.554733074482215]
This paper introduces an innovative tooth segmentation network that integrates boundary information to address the issue of indistinct boundaries between teeth and adjacent tissues.
In the most recent STS Data Challenge, our methodology was rigorously tested and received a commendable overall score of 0.91.
arXiv Detail & Related papers (2024-09-06T02:12:21Z) - Towards a Benchmark for Colorectal Cancer Segmentation in Endorectal Ultrasound Videos: Dataset and Model Development [59.74920439478643]
In this paper, we collect and annotated the first benchmark dataset that covers diverse ERUS scenarios.
Our ERUS-10K dataset comprises 77 videos and 10,000 high-resolution annotated frames.
We introduce a benchmark model for colorectal cancer segmentation, named the Adaptive Sparse-context TRansformer (ASTR)
arXiv Detail & Related papers (2024-08-19T15:04:42Z) - Instance Segmentation and Teeth Classification in Panoramic X-rays [35.8246552579468]
Teeth segmentation and recognition are critical in various dental applications and dental diagnosis.
This article offers a pipeline of two deep learning models, U-Net and YOLOv8, which results in BB-UNet.
We have improved the quality and reliability of teeth segmentation by utilising the YOLOv8 and U-Net capabilities.
arXiv Detail & Related papers (2024-06-06T04:57:29Z) - Multiclass Segmentation using Teeth Attention Modules for Dental X-ray
Images [8.041659727964305]
We propose a novel teeth segmentation model incorporating an M-Net-like structure with Swin Transformers and TAB.
The proposed TAB utilizes a unique attention mechanism that focuses specifically on the complex structures of teeth.
The proposed architecture effectively captures local and global contextual information, accurately defining each tooth and its surrounding structures.
arXiv Detail & Related papers (2023-11-07T06:20:34Z) - Towards Unifying Anatomy Segmentation: Automated Generation of a
Full-body CT Dataset via Knowledge Aggregation and Anatomical Guidelines [113.08940153125616]
We generate a dataset of whole-body CT scans with $142$ voxel-level labels for 533 volumes providing comprehensive anatomical coverage.
Our proposed procedure does not rely on manual annotation during the label aggregation stage.
We release our trained unified anatomical segmentation model capable of predicting $142$ anatomical structures on CT data.
arXiv Detail & Related papers (2023-07-25T09:48:13Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - OdontoAI: A human-in-the-loop labeled data set and an online platform to
boost research on dental panoramic radiographs [53.67409169790872]
This study addresses the construction of a public data set of dental panoramic radiographs.
We benefit from the human-in-the-loop (HITL) concept to expedite the labeling procedure.
Results demonstrate a 51% labeling time reduction using HITL, saving us more than 390 continuous working hours.
arXiv Detail & Related papers (2022-03-29T18:57:23Z) - AI-enabled Automatic Multimodal Fusion of Cone-Beam CT and Intraoral
Scans for Intelligent 3D Tooth-Bone Reconstruction and Clinical Applications [29.065668174732014]
A critical step in virtual dental treatment planning is to accurately delineate all tooth-bone structures from CBCT.
Previous studies have established several methods for CBCT segmentation using deep learning.
Here, we present a Deep Dental Multimodal Analysis framework consisting of a CBCT segmentation model, an intraoral scan (IOS) segmentation model, and a fusion model to generate 3D fused crown-root-bone structures.
arXiv Detail & Related papers (2022-03-11T07:50:15Z) - Improving Classification Model Performance on Chest X-Rays through Lung
Segmentation [63.45024974079371]
We propose a deep learning approach to enhance abnormal chest x-ray (CXR) identification performance through segmentations.
Our approach is designed in a cascaded manner and incorporates two modules: a deep neural network with criss-cross attention modules (XLSor) for localizing lung region in CXR images and a CXR classification model with a backbone of a self-supervised momentum contrast (MoCo) model pre-trained on large-scale CXR data sets.
arXiv Detail & Related papers (2022-02-22T15:24:06Z) - Two-Stage Mesh Deep Learning for Automated Tooth Segmentation and
Landmark Localization on 3D Intraoral Scans [56.55092443401416]
emphiMeshSegNet in the first stage of TS-MDL reached an averaged Dice similarity coefficient (DSC) at 0.953pm0.076$, significantly outperforming the original MeshSegNet.
PointNet-Reg achieved a mean absolute error (MAE) of $0.623pm0.718, mm$ in distances between the prediction and ground truth for $44$ landmarks, which is superior compared with other networks for landmark detection.
arXiv Detail & Related papers (2021-09-24T13:00:26Z) - Pose-Aware Instance Segmentation Framework from Cone Beam CT Images for
Tooth Segmentation [9.880428545498662]
Individual tooth segmentation from cone beam computed tomography (CBCT) images is essential for an anatomical understanding of orthodontic structures.
The presence of severe metal artifacts in CBCT images hinders the accurate segmentation of each individual tooth.
We propose a neural network for pixel-wise labeling to exploit an instance segmentation framework that is robust to metal artifacts.
arXiv Detail & Related papers (2020-02-06T07:57:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.