Gesture Recognition in Robotic Surgery: a Review
- URL: http://arxiv.org/abs/2102.00027v1
- Date: Fri, 29 Jan 2021 19:13:13 GMT
- Title: Gesture Recognition in Robotic Surgery: a Review
- Authors: Beatrice van Amsterdam, Matthew J. Clarkson, Danail Stoyanov
- Abstract summary: This paper reviews the state-of-the-art in methods for automatic recognition of fine-grained gestures in robotic surgery.
The research field is showing rapid expansion, with the majority of articles published in the last 4 years.
- Score: 16.322875532481326
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Objective: Surgical activity recognition is a fundamental step in
computer-assisted interventions. This paper reviews the state-of-the-art in
methods for automatic recognition of fine-grained gestures in robotic surgery
focusing on recent data-driven approaches and outlines the open questions and
future research directions. Methods: An article search was performed on 5
bibliographic databases with the following search terms: robotic,
robot-assisted, JIGSAWS, surgery, surgical, gesture, fine-grained, surgeme,
action, trajectory, segmentation, recognition, parsing. Selected articles were
classified based on the level of supervision required for training and divided
into different groups representing major frameworks for time series analysis
and data modelling. Results: A total of 52 articles were reviewed. The research
field is showing rapid expansion, with the majority of articles published in
the last 4 years. Deep-learning-based temporal models with discriminative
feature extraction and multi-modal data integration have demonstrated promising
results on small surgical datasets. Currently, unsupervised methods perform
significantly less well than the supervised approaches. Conclusion: The
development of large and diverse open-source datasets of annotated
demonstrations is essential for development and validation of robust solutions
for surgical gesture recognition. While new strategies for discriminative
feature extraction and knowledge transfer, or unsupervised and semi-supervised
approaches, can mitigate the need for data and labels, they have not yet been
demonstrated to achieve comparable performance. Important future research
directions include detection and forecast of gesture-specific errors and
anomalies. Significance: This paper is a comprehensive and structured analysis
of surgical gesture recognition methods aiming to summarize the status of this
rapidly evolving field.
Related papers
- SAR-RARP50: Segmentation of surgical instrumentation and Action
Recognition on Robot-Assisted Radical Prostatectomy Challenge [72.97934765570069]
We release the first multimodal, publicly available, in-vivo, dataset for surgical action recognition and semantic instrumentation segmentation, containing 50 suturing video segments of Robotic Assisted Radical Prostatectomy (RARP)
The aim of the challenge is to enable researchers to leverage the scale of the provided dataset and develop robust and highly accurate single-task action recognition and tool segmentation approaches in the surgical domain.
A total of 12 teams participated in the challenge, contributing 7 action recognition methods, 9 instrument segmentation techniques, and 4 multitask approaches that integrated both action recognition and instrument segmentation.
arXiv Detail & Related papers (2023-12-31T13:32:18Z) - Jumpstarting Surgical Computer Vision [2.7396997668655163]
We employ self-supervised learning to flexibly leverage diverse surgical datasets.
We study phase recognition and critical view of safety in laparoscopic cholecystectomy and laparoscopic hysterectomy.
The composition of pre-training datasets can severely affect the effectiveness of SSL methods for various downstream tasks.
arXiv Detail & Related papers (2023-12-10T18:54:16Z) - Methods and datasets for segmentation of minimally invasive surgical
instruments in endoscopic images and videos: A review of the state of the art [0.0]
We identify and characterize datasets used for method development and evaluation.
The paper focuses on methods that work purely visually, without markers of any kind attached to the instruments.
A discussion of the reviewed literature is provided, highlighting existing shortcomings and emphasizing the available potential for future developments.
arXiv Detail & Related papers (2023-04-25T17:38:41Z) - Nuclei & Glands Instance Segmentation in Histology Images: A Narrative
Review [0.5893124686141781]
Instance segmentation of nuclei and glands in the histology images is an important step in computational pathology workflow.
With the advent of modern hardware, the recent availability of large-scale quality public datasets and organized grand challenges have seen a surge in automated methods.
In this survey, 126 papers illustrating the AI based methods for nuclei and glands segmentation published in the last five years-2022) are deeply analyzed.
arXiv Detail & Related papers (2022-08-26T06:52:15Z) - Dissecting Self-Supervised Learning Methods for Surgical Computer Vision [51.370873913181605]
Self-Supervised Learning (SSL) methods have begun to gain traction in the general computer vision community.
The effectiveness of SSL methods in more complex and impactful domains, such as medicine and surgery, remains limited and unexplored.
We present an extensive analysis of the performance of these methods on the Cholec80 dataset for two fundamental and popular tasks in surgical context understanding, phase recognition and tool presence detection.
arXiv Detail & Related papers (2022-07-01T14:17:11Z) - CholecTriplet2021: A benchmark challenge for surgical action triplet
recognition [66.51610049869393]
This paper presents CholecTriplet 2021: an endoscopic vision challenge organized at MICCAI 2021 for the recognition of surgical action triplets in laparoscopic videos.
We present the challenge setup and assessment of the state-of-the-art deep learning methods proposed by the participants during the challenge.
A total of 4 baseline methods and 19 new deep learning algorithms are presented to recognize surgical action triplets directly from surgical videos, achieving mean average precision (mAP) ranging from 4.2% to 38.1%.
arXiv Detail & Related papers (2022-04-10T18:51:55Z) - Co-Generation and Segmentation for Generalized Surgical Instrument
Segmentation on Unlabelled Data [49.419268399590045]
Surgical instrument segmentation for robot-assisted surgery is needed for accurate instrument tracking and augmented reality overlays.
Deep learning-based methods have shown state-of-the-art performance for surgical instrument segmentation, but their results depend on labelled data.
In this paper, we demonstrate the limited generalizability of these methods on different datasets, including human robot-assisted surgeries.
arXiv Detail & Related papers (2021-03-16T18:41:18Z) - Relational Graph Learning on Visual and Kinematics Embeddings for
Accurate Gesture Recognition in Robotic Surgery [84.73764603474413]
We propose a novel online approach of multi-modal graph network (i.e., MRG-Net) to dynamically integrate visual and kinematics information.
The effectiveness of our method is demonstrated with state-of-the-art results on the public JIGSAWS dataset.
arXiv Detail & Related papers (2020-11-03T11:00:10Z) - Robust Medical Instrument Segmentation Challenge 2019 [56.148440125599905]
Intraoperative tracking of laparoscopic instruments is often a prerequisite for computer and robotic-assisted interventions.
Our challenge was based on a surgical data set comprising 10,040 annotated images acquired from a total of 30 surgical procedures.
The results confirm the initial hypothesis, namely that algorithm performance degrades with an increasing domain gap.
arXiv Detail & Related papers (2020-03-23T14:35:08Z) - Multi-Task Recurrent Neural Network for Surgical Gesture Recognition and
Progress Prediction [17.63619129438996]
We propose a multi-task recurrent neural network for simultaneous recognition of surgical gestures and estimation of a novel formulation of surgical task progress.
We demonstrate that recognition performance improves in multi-task frameworks with progress estimation without any additional manual labelling and training.
arXiv Detail & Related papers (2020-03-10T14:28:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.