Simplify Implant Depth Prediction as Video Grounding: A Texture Perceive Implant Depth Prediction Network
- URL: http://arxiv.org/abs/2406.04603v1
- Date: Fri, 7 Jun 2024 03:24:04 GMT
- Title: Simplify Implant Depth Prediction as Video Grounding: A Texture Perceive Implant Depth Prediction Network
- Authors: Xinquan Yang, Xuguang Li, Xiaoling Luo, Leilei Zeng, Yudi Zhang, Linlin Shen, Yongqiang Deng,
- Abstract summary: We develop a Texture Perceive Implant Depth Prediction Network (TPNet)
TPNet consists of an implant region detector (IRD) and an implant depth prediction network (IDPNet)
experiments on a large dental implant dataset demonstrated that the proposed TPNet achieves superior performance than the existing methods.
- Score: 26.68803827522865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Surgical guide plate is an important tool for the dental implant surgery. However, the design process heavily relies on the dentist to manually simulate the implant angle and depth. When deep neural networks have been applied to assist the dentist quickly locates the implant position, most of them are not able to determine the implant depth. Inspired by the video grounding task which localizes the starting and ending time of the target video segment, in this paper, we simplify the implant depth prediction as video grounding and develop a Texture Perceive Implant Depth Prediction Network (TPNet), which enables us to directly output the implant depth without complex measurements of oral bone. TPNet consists of an implant region detector (IRD) and an implant depth prediction network (IDPNet). IRD is an object detector designed to crop the candidate implant volume from the CBCT, which greatly saves the computation resource. IDPNet takes the cropped CBCT data to predict the implant depth. A Texture Perceive Loss (TPL) is devised to enable the encoder of IDPNet to perceive the texture variation among slices. Extensive experiments on a large dental implant dataset demonstrated that the proposed TPNet achieves superior performance than the existing methods.
Related papers
- TCSloT: Text Guided 3D Context and Slope Aware Triple Network for Dental
Implant Position Prediction [27.020346431680355]
In implant prosthesis treatment, the surgical guide of implant is used to ensure accurate implantation.
Deep neural network has been proposed to assist the dentist in locating the implant position.
In this paper, we design a Text Guided 3D Context and Slope Aware Triple Network (TCSloT)
arXiv Detail & Related papers (2023-08-10T05:51:21Z) - TCEIP: Text Condition Embedded Regression Network for Dental Implant
Position Prediction [28.824994303170882]
Deep neural network has been proposed to assist the dentist in designing the location of dental implant.
literature works do not work well when there are multiple missing teeth and easily generate false predictions when the teeth are sparsely distributed.
We propose a text condition embedded implant position regression network (TCEIP), to embed the text condition into the encoder-decoder framework.
arXiv Detail & Related papers (2023-06-26T03:38:43Z) - Two-Stream Regression Network for Dental Implant Position Prediction [29.864052654136614]
In implant prosthesis treatment, the design of the surgical guide heavily relies on the manual location of the implant position.
Deep learning based methods has started to be applied to address this problem.
The space between teeth are various and some of them might present similar texture characteristic with the actual implant region.
arXiv Detail & Related papers (2023-05-17T08:37:26Z) - ImplantFormer: Vision Transformer based Implant Position Regression
Using Dental CBCT Data [27.020346431680355]
Implant prosthesis is the most appropriate treatment for dentition defect or dentition loss, which usually involves a surgical guide design process to decide the implant position.
In this paper, a transformer-based Implant Position Regression Network, ImplantFormer, is proposed to automatically predict the implant position based on the oral CBCT data.
We creatively propose to predict the implant position using the 2D axial view of the tooth crown area and fit a centerline of the implant to obtain the actual implant position at the tooth root.
arXiv Detail & Related papers (2022-10-29T02:31:27Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - A Single-Target License Plate Detection with Attention [56.83051142257412]
Neural Network is commonly adopted to the License Plate Detection (LPD) task and achieves much better performance and precision, especially CNN-based networks can achieve state of the art RetinaNet.
For a single object detection task such as LPD, modified general object detection would be time-consuming, unable to cope with complex scenarios and a cumbersome weights file that is too hard to deploy on the embedded device.
arXiv Detail & Related papers (2021-12-12T03:00:03Z) - 3DVNet: Multi-View Depth Prediction and Volumetric Refinement [68.68537312256144]
3DVNet is a novel multi-view stereo (MVS) depth-prediction method.
Our key idea is the use of a 3D scene-modeling network that iteratively updates a set of coarse depth predictions.
We show that our method exceeds state-of-the-art accuracy in both depth prediction and 3D reconstruction metrics.
arXiv Detail & Related papers (2021-12-01T00:52:42Z) - Two-Stage Mesh Deep Learning for Automated Tooth Segmentation and
Landmark Localization on 3D Intraoral Scans [56.55092443401416]
emphiMeshSegNet in the first stage of TS-MDL reached an averaged Dice similarity coefficient (DSC) at 0.953pm0.076$, significantly outperforming the original MeshSegNet.
PointNet-Reg achieved a mean absolute error (MAE) of $0.623pm0.718, mm$ in distances between the prediction and ground truth for $44$ landmarks, which is superior compared with other networks for landmark detection.
arXiv Detail & Related papers (2021-09-24T13:00:26Z) - Self-Supervised Generative Adversarial Network for Depth Estimation in
Laparoscopic Images [13.996932179049978]
We propose SADepth, a new self-supervised depth estimation method based on Generative Adversarial Networks.
It consists of an encoder-decoder generator and a discriminator to incorporate geometry constraints during training.
Experiments on two public datasets show that SADepth outperforms recent state-of-the-art unsupervised methods by a large margin.
arXiv Detail & Related papers (2021-07-09T19:40:20Z) - Sparse Auxiliary Networks for Unified Monocular Depth Prediction and
Completion [56.85837052421469]
Estimating scene geometry from data obtained with cost-effective sensors is key for robots and self-driving cars.
In this paper, we study the problem of predicting dense depth from a single RGB image with optional sparse measurements from low-cost active depth sensors.
We introduce Sparse Networks (SANs), a new module enabling monodepth networks to perform both the tasks of depth prediction and completion.
arXiv Detail & Related papers (2021-03-30T21:22:26Z) - PLADE-Net: Towards Pixel-Level Accuracy for Self-Supervised Single-View
Depth Estimation with Neural Positional Encoding and Distilled Matting Loss [49.66736599668501]
We propose a self-supervised single-view pixel-level accurate depth estimation network, called PLADE-Net.
Our method shows unprecedented accuracy levels, exceeding 95% in terms of the $delta1$ metric on the KITTI dataset.
arXiv Detail & Related papers (2021-03-12T15:54:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.