A Novel Augmented Reality Ultrasound Framework Using an RGB-D Camera and
a 3D-printed Marker
- URL: http://arxiv.org/abs/2205.04350v1
- Date: Mon, 9 May 2022 14:54:47 GMT
- Title: A Novel Augmented Reality Ultrasound Framework Using an RGB-D Camera and
a 3D-printed Marker
- Authors: Yitian Zhou, Ga\'etan Lelu, Boris Labb\'e, Guillaume Pasquier, Pierre
Le Gargasson, Albert Murienne and Laurent Launay
- Abstract summary: Our goal is to develop a simple and low cost augmented reality echography framework using a standard RGB-D Camera.
Prototype system consisted of an Occipital Structure Core RGB-D camera, a specifically-designed 3D marker, and a fast point cloud registration algorithm FaVoR.
Prototype probe was calibrated on a 3D-printed N-wire phantom using the software PLUS toolkit.
- Score: 0.3061098887924466
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Purpose. Ability to locate and track ultrasound images in the 3D operating
space is of great benefit for multiple clinical applications. This is often
accomplished by tracking the probe using a precise but expensive optical or
electromagnetic tracking system. Our goal is to develop a simple and low cost
augmented reality echography framework using a standard RGB-D Camera.
Methods. A prototype system consisting of an Occipital Structure Core RGB-D
camera, a specifically-designed 3D marker, and a fast point cloud registration
algorithm FaVoR was developed and evaluated on an Ultrasonix ultrasound system.
The probe was calibrated on a 3D-printed N-wire phantom using the software PLUS
toolkit. The proposed calibration method is simplified, requiring no additional
markers or sensors attached to the phantom. Also, a visualization software
based on OpenGL was developed for the augmented reality application.
Results. The calibrated probe was used to augment a real-world video in a
simulated needle insertion scenario. The ultrasound images were rendered on the
video, and visually-coherent results were observed. We evaluated the end-to-end
accuracy of our AR US framework on localizing a cube of 5 cm size. From our two
experiments, the target pose localization error ranges from 5.6 to 5.9 mm and
from -3.9 to 4.2 degrees.
Conclusion. We believe that with the potential democratization of RGB-D
cameras integrated in mobile devices and AR glasses in the future, our
prototype solution may facilitate the use of 3D freehand ultrasound in clinical
routine. Future work should include a more rigorous and thorough evaluation, by
comparing the calibration accuracy with those obtained by commercial tracking
solutions in both simulated and real medical scenarios.
Related papers
- Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - MM3DGS SLAM: Multi-modal 3D Gaussian Splatting for SLAM Using Vision, Depth, and Inertial Measurements [59.70107451308687]
We show for the first time that using 3D Gaussians for map representation with unposed camera images and inertial measurements can enable accurate SLAM.
Our method, MM3DGS, addresses the limitations of prior rendering by enabling faster scale awareness, and improved trajectory tracking.
We also release a multi-modal dataset, UT-MM, collected from a mobile robot equipped with a camera and an inertial measurement unit.
arXiv Detail & Related papers (2024-04-01T04:57:41Z) - Real-Time Radiance Fields for Single-Image Portrait View Synthesis [85.32826349697972]
We present a one-shot method to infer and render a 3D representation from a single unposed image in real-time.
Given a single RGB input, our image encoder directly predicts a canonical triplane representation of a neural radiance field for 3D-aware novel view synthesis via volume rendering.
Our method is fast (24 fps) on consumer hardware, and produces higher quality results than strong GAN-inversion baselines that require test-time optimization.
arXiv Detail & Related papers (2023-05-03T17:56:01Z) - Freehand 2D Ultrasound Probe Calibration for Image Fusion with 3D MRI/CT [13.4304925464326]
This will enable us to visualize US image data during surgical procedures using augmented reality.
A near-millimeter accuracy can be achieved with the proposed approach.
arXiv Detail & Related papers (2023-03-14T08:55:24Z) - Simple and Effective Synthesis of Indoor 3D Scenes [78.95697556834536]
We study the problem of immersive 3D indoor scenes from one or more images.
Our aim is to generate high-resolution images and videos from novel viewpoints.
We propose an image-to-image GAN that maps directly from reprojections of incomplete point clouds to full high-resolution RGB-D images.
arXiv Detail & Related papers (2022-04-06T17:54:46Z) - Occlusion-robust Visual Markerless Bone Tracking for Computer-Assisted
Orthopaedic Surgery [41.681134859412246]
We propose a RGB-D sensing-based markerless tracking method that is robust against occlusion.
By using a high-quality commercial RGB-D camera, our proposed visual tracking method achieves an accuracy of 1-2 degress and 2-4 mm on a model knee.
arXiv Detail & Related papers (2021-08-24T09:49:08Z) - Progressive Coordinate Transforms for Monocular 3D Object Detection [52.00071336733109]
We propose a novel and lightweight approach, dubbed em Progressive Coordinate Transforms (PCT) to facilitate learning coordinate representations.
In this paper, we propose a novel and lightweight approach, dubbed em Progressive Coordinate Transforms (PCT) to facilitate learning coordinate representations.
arXiv Detail & Related papers (2021-08-12T15:22:33Z) - 3D Reconstruction and Alignment by Consumer RGB-D Sensors and Fiducial
Planar Markers for Patient Positioning in Radiation Therapy [1.7744342894757368]
This paper proposes a fast and cheap patient positioning method based on inexpensive consumer level RGB-D sensors.
The proposed method relies on a 3D reconstruction approach that fuses, in real-time, artificial and natural visual landmarks recorded from a hand-held RGB-D sensor.
arXiv Detail & Related papers (2021-03-22T20:20:59Z) - Sensorless Freehand 3D Ultrasound Reconstruction via Deep Contextual
Learning [13.844630500061378]
Current methods for 3D volume reconstruction from freehand US scans require external tracking devices to provide spatial position for every frame.
We propose a deep contextual learning network (DCL-Net), which can efficiently exploit the image feature relationship between US frames and reconstruct 3D US volumes without any tracking device.
arXiv Detail & Related papers (2020-06-13T18:37:30Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.