More Than Meets the Eye: Analyzing Anesthesiologists' Visual Attention
in the Operating Room Using Deep Learning Models
- URL: http://arxiv.org/abs/2308.05501v1
- Date: Thu, 10 Aug 2023 11:12:04 GMT
- Title: More Than Meets the Eye: Analyzing Anesthesiologists' Visual Attention
in the Operating Room Using Deep Learning Models
- Authors: Sapir Gershov, Fadi Mahameed, Aeyal Raz, Shlomi Laufer
- Abstract summary: Currently, most studies employ wearable eye-tracking technologies to analyze anesthesiologists' visual patterns.
By utilizing a novel eye-tracking method in the form of deep learning models that process monitor-mounted webcams, we collected continuous behavioral data.
We distinguished between baseline VA distribution during uneventful periods to patterns associated with active phases or during critical, unanticipated incidents.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Patient's vital signs, which are displayed on monitors, make the
anesthesiologist's visual attention (VA) a key component in the safe management
of patients under general anesthesia; moreover, the distribution of said VA and
the ability to acquire specific cues throughout the anesthetic, may have a
direct impact on patient's outcome. Currently, most studies employ wearable
eye-tracking technologies to analyze anesthesiologists' visual patterns. Albeit
being able to produce meticulous data, wearable devices are not a sustainable
solution for large-scale or long-term use for data collection in the operating
room (OR). Thus, by utilizing a novel eye-tracking method in the form of deep
learning models that process monitor-mounted webcams, we collected continuous
behavioral data and gained insight into the anesthesiologist's VA distribution
with minimal disturbance to their natural workflow. In this study, we collected
OR video recordings using the proposed framework and compared different visual
behavioral patterns. We distinguished between baseline VA distribution during
uneventful periods to patterns associated with active phases or during
critical, unanticipated incidents. In the future, such a platform may serve as
a crucial component of context-aware assistive technologies in the OR.
Related papers
- Dr-LLaVA: Visual Instruction Tuning with Symbolic Clinical Grounding [53.629132242389716]
Vision-Language Models (VLM) can support clinicians by analyzing medical images and engaging in natural language interactions.
VLMs often exhibit "hallucinogenic" behavior, generating textual outputs not grounded in contextual multimodal information.
We propose a new alignment algorithm that uses symbolic representations of clinical reasoning to ground VLMs in medical knowledge.
arXiv Detail & Related papers (2024-05-29T23:19:28Z) - CathFlow: Self-Supervised Segmentation of Catheters in Interventional Ultrasound Using Optical Flow and Transformers [66.15847237150909]
We introduce a self-supervised deep learning architecture to segment catheters in longitudinal ultrasound images.
The network architecture builds upon AiAReSeg, a segmentation transformer built with the Attention in Attention mechanism.
We validated our model on a test dataset, consisting of unseen synthetic data and images collected from silicon aorta phantoms.
arXiv Detail & Related papers (2024-03-21T15:13:36Z) - Latent Space Explorer: Visual Analytics for Multimodal Latent Space
Exploration [12.202104074544202]
A multimodal machine learning model trained from large datasets can potentially predict the onset of heart-related diseases.
Latent Space Explorer provides interactive visualizations that enable users to explore the multimodal representation of subjects.
A user study was conducted with medical experts and their feedback provided useful insights into how Latent Space Explorer can help their analysis.
arXiv Detail & Related papers (2023-12-01T15:25:56Z) - Deep denoising autoencoder-based non-invasive blood flow detection for
arteriovenous fistula [10.030431512848239]
We propose an approach based on deep denoising autoencoders (DAEs) that perform dimensionality reduction and reconstruction tasks.
Our results demonstrate that the latent representation generated by the DAE surpasses expectations with an accuracy of 0.93.
The incorporation of noise-mixing and the utilization of a noise-to-clean scheme effectively enhance the discriminative capabilities of the latent representation.
arXiv Detail & Related papers (2023-06-12T04:46:01Z) - Safe AI for health and beyond -- Monitoring to transform a health
service [51.8524501805308]
We will assess the infrastructure required to monitor the outputs of a machine learning algorithm.
We will present two scenarios with examples of monitoring and updates of models.
arXiv Detail & Related papers (2023-03-02T17:27:45Z) - A Deep Learning Approach for the Segmentation of Electroencephalography
Data in Eye Tracking Applications [56.458448869572294]
We introduce DETRtime, a novel framework for time-series segmentation of EEG data.
Our end-to-end deep learning-based framework brings advances in Computer Vision to the forefront.
Our model generalizes well in the task of EEG sleep stage segmentation.
arXiv Detail & Related papers (2022-06-17T10:17:24Z) - A Temporal Learning Approach to Inpainting Endoscopic Specularities and
Its effect on Image Correspondence [13.25903945009516]
We propose using a temporal generative adversarial network (GAN) to inpaint the hidden anatomy under specularities.
This is achieved using in-vivo data of gastric endoscopy (Hyper-Kvasir) in a fully unsupervised manner.
We also assess the effect of our method in computer vision tasks that underpin 3D reconstruction and camera motion estimation.
arXiv Detail & Related papers (2022-03-31T13:14:00Z) - CheXstray: Real-time Multi-Modal Data Concordance for Drift Detection in
Medical Imaging AI [1.359138408203412]
We build and test a medical imaging AI drift monitoring workflow that tracks data and model drift without contemporaneous ground truth.
Key contributions include (1) proof-of-concept for medical imaging drift detection including use of VAE and domain specific statistical methods.
This work has important implications for addressing the translation gap related to continuous medical imaging AI model monitoring in dynamic healthcare environments.
arXiv Detail & Related papers (2022-02-06T18:58:35Z) - In-Bed Human Pose Estimation from Unseen and Privacy-Preserving Image
Domains [22.92165116962952]
In-bed human posture estimation provides important health-related metrics with potential value in medical condition assessments.
We propose a multi-modal conditional variational autoencoder (MC-VAE) capable of reconstructing features from missing modalities seen during training.
We demonstrate that body positions can be effectively recognized from the available modality, achieving on par results with baseline models.
arXiv Detail & Related papers (2021-11-30T04:56:16Z) - On the Robustness of Pretraining and Self-Supervision for a Deep
Learning-based Analysis of Diabetic Retinopathy [70.71457102672545]
We compare the impact of different training procedures for diabetic retinopathy grading.
We investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions.
Our results indicate that models from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions.
arXiv Detail & Related papers (2021-06-25T08:32:45Z) - An Interpretable Multiple-Instance Approach for the Detection of
referable Diabetic Retinopathy from Fundus Images [72.94446225783697]
We propose a machine learning system for the detection of referable Diabetic Retinopathy in fundus images.
By extracting local information from image patches and combining it efficiently through an attention mechanism, our system is able to achieve high classification accuracy.
We evaluate our approach on publicly available retinal image datasets, in which it exhibits near state-of-the-art performance.
arXiv Detail & Related papers (2021-03-02T13:14:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.