Identification of Cognitive Workload during Surgical Tasks with
Multimodal Deep Learning
- URL: http://arxiv.org/abs/2209.06208v1
- Date: Mon, 12 Sep 2022 18:29:34 GMT
- Title: Identification of Cognitive Workload during Surgical Tasks with
Multimodal Deep Learning
- Authors: Kaizhe Jin, Adrian Rubio-Solis, Ravik Nain, Tochukwu Onyeogulu, Amirul
Islam, Salman Khan, Tochukwu Onyeogulu, Amirul Islam, Salman Khan, Izzeddin
Teeti, Fabio Cuzzolin and George Mylonas
- Abstract summary: An increase in the associated Cognitive Workload (CWL) results from dealing with unexpected and repetitive tasks.
In this paper, a cascade of two machine learning approaches is suggested for the multimodal recognition of CWL.
A Convolutional Neural Network (CNN) uses this information to identify different types of CWL associated to each surgical task.
- Score: 20.706268332427157
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In operating Rooms (ORs), activities are usually different from other typical
working environments. In particular, surgeons are frequently exposed to
multiple psycho-organizational constraints that may cause negative
repercussions on their health and performance. This is commonly attributed to
an increase in the associated Cognitive Workload (CWL) that results from
dealing with unexpected and repetitive tasks, as well as large amounts of
information and potentially risky cognitive overload. In this paper, a cascade
of two machine learning approaches is suggested for the multimodal recognition
of CWL in a number of four different surgical tasks. First, a model based on
the concept of transfer learning is used to identify if a surgeon is
experiencing any CWL. Secondly, a Convolutional Neural Network (CNN) uses this
information to identify different types of CWL associated to each surgical
task. The suggested multimodal approach consider adjacent signals from
electroencephalogram (EEG), functional near-infrared spectroscopy (fNIRS) and
pupil eye diameter. The concatenation of signals allows complex correlations in
terms of time (temporal) and channel location (spatial). Data collection is
performed by a Multi-sensing AI Environment for Surgical Task $\&$ Role
Optimisation platform (MAESTRO) developed at HARMS Lab. To compare the
performance of the proposed methodology, a number of state-of-art machine
learning techniques have been implemented. The tests show that the proposed
model has a precision of 93%.
Related papers
- Hypergraph-Transformer (HGT) for Interactive Event Prediction in
Laparoscopic and Robotic Surgery [50.3022015601057]
We propose a predictive neural network that is capable of understanding and predicting critical interactive aspects of surgical workflow from intra-abdominal video.
We verify our approach on established surgical datasets and applications, including the detection and prediction of action triplets.
Our results demonstrate the superiority of our approach compared to unstructured alternatives.
arXiv Detail & Related papers (2024-02-03T00:58:05Z) - ST(OR)2: Spatio-Temporal Object Level Reasoning for Activity Recognition
in the Operating Room [6.132617753806978]
We propose a new sample-efficient and object-based approach for surgical activity recognition in the OR.
Our method focuses on the geometric arrangements between clinicians and surgical devices, thus utilizing the significant object interaction dynamics in the OR.
arXiv Detail & Related papers (2023-12-19T15:33:57Z) - Jumpstarting Surgical Computer Vision [2.7396997668655163]
We employ self-supervised learning to flexibly leverage diverse surgical datasets.
We study phase recognition and critical view of safety in laparoscopic cholecystectomy and laparoscopic hysterectomy.
The composition of pre-training datasets can severely affect the effectiveness of SSL methods for various downstream tasks.
arXiv Detail & Related papers (2023-12-10T18:54:16Z) - Evaluating the structure of cognitive tasks with transfer learning [67.22168759751541]
This study investigates the transferability of deep learning representations between different EEG decoding tasks.
We conduct extensive experiments using state-of-the-art decoding models on two recently released EEG datasets.
arXiv Detail & Related papers (2023-07-28T14:51:09Z) - Dissecting Self-Supervised Learning Methods for Surgical Computer Vision [51.370873913181605]
Self-Supervised Learning (SSL) methods have begun to gain traction in the general computer vision community.
The effectiveness of SSL methods in more complex and impactful domains, such as medicine and surgery, remains limited and unexplored.
We present an extensive analysis of the performance of these methods on the Cholec80 dataset for two fundamental and popular tasks in surgical context understanding, phase recognition and tool presence detection.
arXiv Detail & Related papers (2022-07-01T14:17:11Z) - CholecTriplet2021: A benchmark challenge for surgical action triplet
recognition [66.51610049869393]
This paper presents CholecTriplet 2021: an endoscopic vision challenge organized at MICCAI 2021 for the recognition of surgical action triplets in laparoscopic videos.
We present the challenge setup and assessment of the state-of-the-art deep learning methods proposed by the participants during the challenge.
A total of 4 baseline methods and 19 new deep learning algorithms are presented to recognize surgical action triplets directly from surgical videos, achieving mean average precision (mAP) ranging from 4.2% to 38.1%.
arXiv Detail & Related papers (2022-04-10T18:51:55Z) - Learning Invariant Representation of Tasks for Robust Surgical State
Estimation [39.515036686428836]
We propose StiseNet, a Surgical Task Invariance State Estimation Network.
StiseNet minimizes the effects of variations in surgical technique and operating environments inherent to RAS datasets.
It is shown to outperform state-of-the-art state estimation methods on three datasets.
arXiv Detail & Related papers (2021-02-18T02:32:50Z) - Relational Graph Learning on Visual and Kinematics Embeddings for
Accurate Gesture Recognition in Robotic Surgery [84.73764603474413]
We propose a novel online approach of multi-modal graph network (i.e., MRG-Net) to dynamically integrate visual and kinematics information.
The effectiveness of our method is demonstrated with state-of-the-art results on the public JIGSAWS dataset.
arXiv Detail & Related papers (2020-11-03T11:00:10Z) - Aggregating Long-Term Context for Learning Laparoscopic and
Robot-Assisted Surgical Workflows [40.48632897750319]
We propose a new temporal network structure that leverages task-specific network representation to collect long-term sufficient statistics.
We demonstrate superior results over existing and novel state-of-the-art segmentation techniques on two laparoscopic cholecystectomy datasets.
arXiv Detail & Related papers (2020-09-01T20:29:14Z) - LRTD: Long-Range Temporal Dependency based Active Learning for Surgical
Workflow Recognition [67.86810761677403]
We propose a novel active learning method for cost-effective surgical video analysis.
Specifically, we propose a non-local recurrent convolutional network (NL-RCNet), which introduces non-local block to capture the long-range temporal dependency.
We validate our approach on a large surgical video dataset (Cholec80) by performing surgical workflow recognition task.
arXiv Detail & Related papers (2020-04-21T09:21:22Z) - TeCNO: Surgical Phase Recognition with Multi-Stage Temporal
Convolutional Networks [43.95869213955351]
We propose a Multi-Stage Temporal Convolutional Network (MS-TCN) that performs hierarchical prediction refinement for surgical phase recognition.
Our method is thoroughly evaluated on two datasets of laparoscopic cholecystectomy videos with and without the use of additional surgical tool information.
arXiv Detail & Related papers (2020-03-24T10:12:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.