Spatiotemporal-Aware Augmented Reality: Redefining HCI in Image-Guided
Therapy
- URL: http://arxiv.org/abs/2003.02260v1
- Date: Wed, 4 Mar 2020 18:59:55 GMT
- Title: Spatiotemporal-Aware Augmented Reality: Redefining HCI in Image-Guided
Therapy
- Authors: Javad Fotouhi, Arian Mehrfard, Tianyu Song, Alex Johnson, Greg Osgood,
Mathias Unberath, Mehran Armand, and Nassir Navab
- Abstract summary: Augmented reality (AR) has been introduced in the operating rooms in the last decade.
This paper shows how exemplary visualization are redefined by taking full advantage of head-mounted displays.
The awareness of the system from the geometric and physical characteristics of X-ray imaging allows the redefinition of different human-machine interfaces.
- Score: 39.370739217840594
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Suboptimal interaction with patient data and challenges in mastering 3D
anatomy based on ill-posed 2D interventional images are essential concerns in
image-guided therapies. Augmented reality (AR) has been introduced in the
operating rooms in the last decade; however, in image-guided interventions, it
has often only been considered as a visualization device improving traditional
workflows. As a consequence, the technology is gaining minimum maturity that it
requires to redefine new procedures, user interfaces, and interactions. The
main contribution of this paper is to reveal how exemplary workflows are
redefined by taking full advantage of head-mounted displays when entirely
co-registered with the imaging system at all times. The proposed AR landscape
is enabled by co-localizing the users and the imaging devices via the operating
room environment and exploiting all involved frustums to move spatial
information between different bodies. The awareness of the system from the
geometric and physical characteristics of X-ray imaging allows the redefinition
of different human-machine interfaces. We demonstrate that this AR paradigm is
generic, and can benefit a wide variety of procedures. Our system achieved an
error of $4.76\pm2.91$ mm for placing K-wire in a fracture management
procedure, and yielded errors of $1.57\pm1.16^\circ$ and $1.46\pm1.00^\circ$ in
the abduction and anteversion angles, respectively, for total hip arthroplasty.
We hope that our holistic approach towards improving the interface of surgery
not only augments the surgeon's capabilities but also augments the surgical
team's experience in carrying out an effective intervention with reduced
complications and provide novel approaches of documenting procedures for
training purposes.
Related papers
- Deep intra-operative illumination calibration of hyperspectral cameras [73.08443963791343]
Hyperspectral imaging (HSI) is emerging as a promising novel imaging modality with various potential surgical applications.
We show that dynamically changing lighting conditions in the operating room dramatically affect the performance of HSI applications.
We propose a novel learning-based approach to automatically recalibrating hyperspectral images during surgery.
arXiv Detail & Related papers (2024-09-11T08:30:03Z) - Creating a Digital Twin of Spinal Surgery: A Proof of Concept [68.37190859183663]
Surgery digitalization is the process of creating a virtual replica of real-world surgery.
We present a proof of concept (PoC) for surgery digitalization that is applied to an ex-vivo spinal surgery.
We employ five RGB-D cameras for dynamic 3D reconstruction of the surgeon, a high-end camera for 3D reconstruction of the anatomy, an infrared stereo camera for surgical instrument tracking, and a laser scanner for 3D reconstruction of the operating room and data fusion.
arXiv Detail & Related papers (2024-03-25T13:09:40Z) - SLIMBRAIN: Augmented Reality Real-Time Acquisition and Processing System For Hyperspectral Classification Mapping with Depth Information for In-Vivo Surgical Procedures [0.0]
This paper presents SLIMBRAIN, a real-time acquisition and processing AR system suitable to classify and display brain tumor tissue from hyperspectral (HS) information.
The system captures and processes HS images at 14 frames per second (FPS) during the course of a tumor resection operation to detect and delimit cancer tissue at the same time the neurosurgeon operates.
The result is represented in an AR visualization where the classification results are overlapped with the RGB point cloud captured by a LiDAR camera.
arXiv Detail & Related papers (2024-03-25T11:10:49Z) - Monocular Microscope to CT Registration using Pose Estimation of the
Incus for Augmented Reality Cochlear Implant Surgery [3.8909273404657556]
We develop a method that permits direct 2D-to-3D registration of the view microscope video to the pre-operative Computed Tomography (CT) scan without the need for external tracking equipment.
Our results demonstrate the accuracy with an average rotation error of less than 25 degrees and a translation error of less than 2 mm, 3 mm, and 0.55% for the x, y, and z axes, respectively.
arXiv Detail & Related papers (2024-03-12T00:26:08Z) - Surgical Temporal Action-aware Network with Sequence Regularization for
Phase Recognition [28.52533700429284]
We propose a Surgical Temporal Action-aware Network with sequence Regularization, named STAR-Net, to recognize surgical phases more accurately from input videos.
MS-STA module integrates visual features with spatial and temporal knowledge of surgical actions at the cost of 2D networks.
Our STAR-Net with MS-STA and DSR can exploit visual features of surgical actions with effective regularization, thereby leading to the superior performance of surgical phase recognition.
arXiv Detail & Related papers (2023-11-21T13:43:16Z) - End-to-End assessment of AR-assisted neurosurgery systems [0.5892638927736115]
We classify different techniques for assessing an AR-assisted neurosurgery system and propose a new technique to systematize the assessment procedure.
We found that although the system can undergo registration and tracking errors, physical feedback can significantly reduce the error caused by hologram displacement.
The lack of visual feedback on the hologram does not have a significant effect on the user 3D perception.
arXiv Detail & Related papers (2023-11-03T13:41:44Z) - Visual-Kinematics Graph Learning for Procedure-agnostic Instrument Tip
Segmentation in Robotic Surgeries [29.201385352740555]
We propose a novel visual-kinematics graph learning framework to accurately segment the instrument tip given various surgical procedures.
Specifically, a graph learning framework is proposed to encode relational features of instrument parts from both image and kinematics.
A cross-modal contrastive loss is designed to incorporate robust geometric prior from kinematics to image for tip segmentation.
arXiv Detail & Related papers (2023-09-02T14:52:58Z) - GLSFormer : Gated - Long, Short Sequence Transformer for Step
Recognition in Surgical Videos [57.93194315839009]
We propose a vision transformer-based approach to learn temporal features directly from sequence-level patches.
We extensively evaluate our approach on two cataract surgery video datasets, Cataract-101 and D99, and demonstrate superior performance compared to various state-of-the-art methods.
arXiv Detail & Related papers (2023-07-20T17:57:04Z) - Next-generation Surgical Navigation: Marker-less Multi-view 6DoF Pose
Estimation of Surgical Instruments [66.74633676595889]
We present a multi-camera capture setup consisting of static and head-mounted cameras.
Second, we publish a multi-view RGB-D video dataset of ex-vivo spine surgeries, captured in a surgical wet lab and a real operating theatre.
Third, we evaluate three state-of-the-art single-view and multi-view methods for the task of 6DoF pose estimation of surgical instruments.
arXiv Detail & Related papers (2023-05-05T13:42:19Z) - Multimodal Semantic Scene Graphs for Holistic Modeling of Surgical
Procedures [70.69948035469467]
We take advantage of the latest computer vision methodologies for generating 3D graphs from camera views.
We then introduce the Multimodal Semantic Graph Scene (MSSG) which aims at providing unified symbolic and semantic representation of surgical procedures.
arXiv Detail & Related papers (2021-06-09T14:35:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.