SuPRA: Surgical Phase Recognition and Anticipation for Intra-Operative
Planning
- URL: http://arxiv.org/abs/2403.06200v1
- Date: Sun, 10 Mar 2024 12:46:33 GMT
- Title: SuPRA: Surgical Phase Recognition and Anticipation for Intra-Operative
Planning
- Authors: Maxence Boels, Yang Liu, Prokar Dasgupta, Alejandro Granados,
Sebastien Ourselin
- Abstract summary: We propose a dual approach that simultaneously recognises the current surgical phase and predicts upcoming ones.
Our novel method, Surgical Phase Recognition and Anticipation (SuPRA), leverages past and current information for accurate intra-operative phase recognition.
- Score: 46.57714869178571
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Intra-operative recognition of surgical phases holds significant potential
for enhancing real-time contextual awareness in the operating room. However, we
argue that online recognition, while beneficial, primarily lends itself to
post-operative video analysis due to its limited direct impact on the actual
surgical decisions and actions during ongoing procedures. In contrast, we
contend that the prediction and anticipation of surgical phases are inherently
more valuable for intra-operative assistance, as they can meaningfully
influence a surgeon's immediate and long-term planning by providing foresight
into future steps. To address this gap, we propose a dual approach that
simultaneously recognises the current surgical phase and predicts upcoming
ones, thus offering comprehensive intra-operative assistance and guidance on
the expected remaining workflow. Our novel method, Surgical Phase Recognition
and Anticipation (SuPRA), leverages past and current information for accurate
intra-operative phase recognition while using future segments for phase
prediction. This unified approach challenges conventional frameworks that treat
these objectives separately. We have validated SuPRA on two reputed datasets,
Cholec80 and AutoLaparo21, where it demonstrated state-of-the-art performance
with recognition accuracies of 91.8% and 79.3%, respectively. Additionally, we
introduce and evaluate our model using new segment-level evaluation metrics,
namely Edit and F1 Overlap scores, for a more temporal assessment of segment
classification. In conclusion, SuPRA presents a new multi-task approach that
paves the way for improved intra-operative assistance through surgical phase
recognition and prediction of future events.
Related papers
- Peri-AIIMS: Perioperative Artificial Intelligence Driven Integrated Modeling of Surgeries using Anesthetic, Physical and Cognitive Statuses for Predicting Hospital Outcomes [12.493423568689801]
preoperative cognitive status as measured by the clock drawing test contributed to predicting length of hospital stay, hospital charges, average pain experienced during follow-up.
Machine learning models were trained to classify postoperative outcomes in hold-out test sets.
arXiv Detail & Related papers (2024-10-29T23:42:51Z) - Robust Surgical Phase Recognition From Annotation Efficient Supervision [1.1510009152620668]
We propose a robust method for surgical phase recognition that can handle missing phase annotations effectively.
We achieve an accuracy of 85.1% on the MultiBypass140 dataset using only 3 annotated frames per video.
Our work contributes to the advancement of surgical workflow recognition and paves the way for more efficient and reliable surgical phase recognition systems.
arXiv Detail & Related papers (2024-06-26T16:47:31Z) - Hypergraph-Transformer (HGT) for Interactive Event Prediction in
Laparoscopic and Robotic Surgery [50.3022015601057]
We propose a predictive neural network that is capable of understanding and predicting critical interactive aspects of surgical workflow from intra-abdominal video.
We verify our approach on established surgical datasets and applications, including the detection and prediction of action triplets.
Our results demonstrate the superiority of our approach compared to unstructured alternatives.
arXiv Detail & Related papers (2024-02-03T00:58:05Z) - GLSFormer : Gated - Long, Short Sequence Transformer for Step
Recognition in Surgical Videos [57.93194315839009]
We propose a vision transformer-based approach to learn temporal features directly from sequence-level patches.
We extensively evaluate our approach on two cataract surgery video datasets, Cataract-101 and D99, and demonstrate superior performance compared to various state-of-the-art methods.
arXiv Detail & Related papers (2023-07-20T17:57:04Z) - LoViT: Long Video Transformer for Surgical Phase Recognition [59.06812739441785]
We present a two-stage method, called Long Video Transformer (LoViT) for fusing short- and long-term temporal information.
Our approach outperforms state-of-the-art methods on the Cholec80 and AutoLaparo datasets consistently.
arXiv Detail & Related papers (2023-05-15T20:06:14Z) - Towards Holistic Surgical Scene Understanding [1.004785607987398]
We present a new experimental framework towards holistic surgical scene understanding.
First, we introduce the Phase, Step, Instrument, and Atomic Visual Action recognition (PSI-AVA) dataset.
Second, we present Transformers for Action, Phase, Instrument, and steps Recognition (TAPIR) as a strong baseline for surgical scene understanding.
arXiv Detail & Related papers (2022-12-08T22:15:27Z) - CholecTriplet2021: A benchmark challenge for surgical action triplet
recognition [66.51610049869393]
This paper presents CholecTriplet 2021: an endoscopic vision challenge organized at MICCAI 2021 for the recognition of surgical action triplets in laparoscopic videos.
We present the challenge setup and assessment of the state-of-the-art deep learning methods proposed by the participants during the challenge.
A total of 4 baseline methods and 19 new deep learning algorithms are presented to recognize surgical action triplets directly from surgical videos, achieving mean average precision (mAP) ranging from 4.2% to 38.1%.
arXiv Detail & Related papers (2022-04-10T18:51:55Z) - SUrgical PRediction GAN for Events Anticipation [38.65189355224683]
We used a novel GAN formulation that sampled the future surgical phases trajectory conditioned, on past laparoscopic video frames.
We demonstrated its effectiveness in inferring and predicting the progress of laparoscopic cholecystectomy videos.
We surveyed surgeons to evaluate the plausibility of these predicted trajectories.
arXiv Detail & Related papers (2021-05-10T19:56:45Z) - TeCNO: Surgical Phase Recognition with Multi-Stage Temporal
Convolutional Networks [43.95869213955351]
We propose a Multi-Stage Temporal Convolutional Network (MS-TCN) that performs hierarchical prediction refinement for surgical phase recognition.
Our method is thoroughly evaluated on two datasets of laparoscopic cholecystectomy videos with and without the use of additional surgical tool information.
arXiv Detail & Related papers (2020-03-24T10:12:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.