Robust Self-Supervised Learning of Deterministic Errors in Single-Plane
(Monoplanar) and Dual-Plane (Biplanar) X-ray Fluoroscopy
- URL: http://arxiv.org/abs/2001.00686v1
- Date: Fri, 3 Jan 2020 01:56:21 GMT
- Title: Robust Self-Supervised Learning of Deterministic Errors in Single-Plane
(Monoplanar) and Dual-Plane (Biplanar) X-ray Fluoroscopy
- Authors: Jacky C.K. Chow, Steven K. Boyd, Derek D. Lichti and Janet L. Ronsky
- Abstract summary: Fluoroscopic imaging that captures X-ray images at video framerates is advantageous for guiding catheter insertions by vascular surgeons and interventional radiologists.
Visualizing the dynamical movements non-invasively allows complex surgical procedures to be performed with less trauma to the patient.
This paper presents a robust self-calibration algorithm suitable for single-plane and dual-plane fluoroscopy.
- Score: 2.7528170226206443
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fluoroscopic imaging that captures X-ray images at video framerates is
advantageous for guiding catheter insertions by vascular surgeons and
interventional radiologists. Visualizing the dynamical movements non-invasively
allows complex surgical procedures to be performed with less trauma to the
patient. To improve surgical precision, endovascular procedures can benefit
from more accurate fluoroscopy data via calibration. This paper presents a
robust self-calibration algorithm suitable for single-plane and dual-plane
fluoroscopy. A three-dimensional (3D) target field was imaged by the
fluoroscope in a strong geometric network configuration. The unknown 3D
positions of targets and the fluoroscope pose were estimated simultaneously by
maximizing the likelihood of the Student-t probability distribution function. A
smoothed k-nearest neighbour (kNN) regression is then used to model the
deterministic component of the image reprojection error of the robust bundle
adjustment. The Maximum Likelihood Estimation step and the kNN regression step
are then repeated iteratively until convergence. Four different error modeling
schemes were compared while varying the quantity of training images. It was
found that using a smoothed kNN regression can automatically model the
systematic errors in fluoroscopy with similar accuracy as a human expert using
a small training dataset. When all training images were used, the 3D mapping
error was reduced from 0.61-0.83 mm to 0.04 mm post-calibration (94.2-95.7%
improvement), and the 2D reprojection error was reduced from 1.17-1.31 to
0.20-0.21 pixels (83.2-83.8% improvement). When using biplanar fluoroscopy, the
3D measurement accuracy of the system improved from 0.60 mm to 0.32 mm (47.2%
improvement).
Related papers
- Deep-Motion-Net: GNN-based volumetric organ shape reconstruction from single-view 2D projections [1.8189671456038365]
We propose an end-to-end graph neural network architecture that enables 3D organ shape reconstruction during radiotherapy.
The proposed model learns the mesh regression from a patient-specific template and deep features extracted from kV images at arbitrary projection angles.
Overall framework was tested quantitatively on synthetic respiratory motion scenarios and qualitatively on in-treatment images acquired over full scan series for liver cancer patients.
arXiv Detail & Related papers (2024-07-09T09:07:18Z) - Intraoperative 2D/3D Image Registration via Differentiable X-ray Rendering [5.617649111108429]
We present DiffPose, a self-supervised approach that leverages patient-specific simulation and differentiable physics-based rendering to achieve accurate 2D/3D registration without relying on manually labeled data.
DiffPose achieves sub-millimeter accuracy across surgical datasets at intraoperative speeds, improving upon existing unsupervised methods by an order of magnitude and even outperforming supervised baselines.
arXiv Detail & Related papers (2023-12-11T13:05:54Z) - X-Ray to CT Rigid Registration Using Scene Coordinate Regression [1.1687067206676627]
This paper proposes a fully automatic registration method that is robust to extreme viewpoints.
It is based on a fully convolutional neural network (CNN) that regresses the overlapping coordinates for a given X-ray image.
The proposed method achieved an average mean target registration error (mTRE) of 3.79 mm in the 50th percentile of the simulated test dataset and projected mTRE of 9.65 mm in the 50th percentile of real fluoroscopic images for pelvis registration.
arXiv Detail & Related papers (2023-11-25T17:48:46Z) - The Impact of Loss Functions and Scene Representations for 3D/2D
Registration on Single-view Fluoroscopic X-ray Pose Estimation [1.758213853394712]
We first develop a differentiable projection rendering framework for the efficient computation of Digitally Reconstructed Radiographs (DRRs)
We then perform pose estimation by iterative descent using various candidate loss functions, that quantify the image discrepancy of the synthesized DRR with respect to the ground-truth fluoroscopic X-ray image.
Using the Mutual Information loss, a comprehensive evaluation of pose estimation performed on a tomographic X-ray dataset of 50 patients$'$ skulls shows that utilizing either discretized (CBCT) or neural (NeTT/mNeRF) scene representations in DiffProj leads to
arXiv Detail & Related papers (2023-08-01T01:12:29Z) - A Spatiotemporal Model for Precise and Efficient Fully-automatic 3D
Motion Correction in OCT [10.550562752812894]
OCT instruments image by-scanning a focused light spot across the retina, acquiring cross-sectional images to generate data.
Patient eye motion during the acquisition poses unique challenges: non-rigid, distorted distortions occur, leading to gaps in data.
We present a new distortion model and a corresponding fully-automatic, reference-free optimization strategy for computational robustness.
arXiv Detail & Related papers (2022-09-15T11:48:53Z) - Uncertainty-Aware Camera Pose Estimation from Points and Lines [101.03675842534415]
Perspective-n-Point-and-Line (Pn$PL) aims at fast, accurate and robust camera localizations with respect to a 3D model from 2D-3D feature coordinates.
arXiv Detail & Related papers (2021-07-08T15:19:36Z) - Robust 360-8PA: Redesigning The Normalized 8-point Algorithm for 360-FoV
Images [53.11097060367591]
We present a novel strategy for estimating an essential matrix from 360-FoV images in spherical projection.
We show that our normalization can increase the camera pose accuracy by about 20% without significantly overhead the time.
arXiv Detail & Related papers (2021-04-22T07:23:11Z) - Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices [48.85784310158493]
We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
arXiv Detail & Related papers (2020-12-16T07:11:16Z) - Probabilistic 3D surface reconstruction from sparse MRI information [58.14653650521129]
We present a novel probabilistic deep learning approach for concurrent 3D surface reconstruction from sparse 2D MR image data and aleatoric uncertainty prediction.
Our method is capable of reconstructing large surface meshes from three quasi-orthogonal MR imaging slices from limited training sets.
arXiv Detail & Related papers (2020-10-05T14:18:52Z) - Cephalometric Landmark Regression with Convolutional Neural Networks on
3D Computed Tomography Data [68.8204255655161]
Cephalometric analysis performed on lateral radiographs doesn't fully exploit the structure of 3D objects due to projection onto the lateral plane.
We present a series of experiments with state of the art 3D convolutional neural network (CNN) based methods for keypoint regression.
For the first time, we extensively evaluate the described methods and demonstrate their effectiveness in the estimation of Frankfort Horizontal and cephalometric points locations.
arXiv Detail & Related papers (2020-07-20T12:45:38Z) - Appearance Learning for Image-based Motion Estimation in Tomography [60.980769164955454]
In tomographic imaging, anatomical structures are reconstructed by applying a pseudo-inverse forward model to acquired signals.
Patient motion corrupts the geometry alignment in the reconstruction process resulting in motion artifacts.
We propose an appearance learning approach recognizing the structures of rigid motion independently from the scanned object.
arXiv Detail & Related papers (2020-06-18T09:49:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.