Surgical tool classification and localization: results and methods from
the MICCAI 2022 SurgToolLoc challenge
- URL: http://arxiv.org/abs/2305.07152v2
- Date: Wed, 31 May 2023 17:17:21 GMT
- Title: Surgical tool classification and localization: results and methods from
the MICCAI 2022 SurgToolLoc challenge
- Authors: Aneeq Zia, Kiran Bhattacharyya, Xi Liu, Max Berniker, Ziheng Wang,
Rogerio Nespolo, Satoshi Kondo, Satoshi Kasai, Kousuke Hirasawa, Bo Liu,
David Austin, Yiheng Wang, Michal Futrega, Jean-Francois Puget, Zhenqiang Li,
Yoichi Sato, Ryo Fujii, Ryo Hachiuma, Mana Masuda, Hideo Saito, An Wang,
Mengya Xu, Mobarakol Islam, Long Bai, Winnie Pang, Hongliang Ren, Chinedu
Nwoye, Luca Sestini, Nicolas Padoy, Maximilian Nielsen, Samuel Sch\"uttler,
Thilo Sentker, H\"umeyra Husseini, Ivo Baltruschat, R\"udiger Schmitz, Ren\'e
Werner, Aleksandr Matsun, Mugariya Farooq, Numan Saaed, Jose Renato Restom
Viera, Mohammad Yaqub, Neil Getty, Fangfang Xia, Zixuan Zhao, Xiaotian Duan,
Xing Yao, Ange Lou, Hao Yang, Jintong Han, Jack Noble, Jie Ying Wu, Tamer
Abdulbaki Alshirbaji, Nour Aldeen Jalal, Herag Arabian, Ning Ding, Knut
Moeller, Weiliang Chen, Quan He, Muhammad Bilal, Taofeek Akinosho, Adnan
Qayyum, Massimo Caputo, Hunaid Vohra, Michael Loizou, Anuoluwapo Ajayi, Ilhem
Berrou, Faatihah Niyi-Odumosu, Lena Maier-Hein, Danail Stoyanov, Stefanie
Speidel, Anthony Jarc
- Abstract summary: We present the results of the SurgLoc 2022 challenge.
The goal was to leverage tool presence data as weak labels for machine learning models trained to detect tools.
We conclude by discussing these results in the broader context of machine learning and surgical data science.
- Score: 69.91670788430162
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The ability to automatically detect and track surgical instruments in
endoscopic videos can enable transformational interventions. Assessing surgical
performance and efficiency, identifying skilled tool use and choreography, and
planning operational and logistical aspects of OR resources are just a few of
the applications that could benefit. Unfortunately, obtaining the annotations
needed to train machine learning models to identify and localize surgical tools
is a difficult task. Annotating bounding boxes frame-by-frame is tedious and
time-consuming, yet large amounts of data with a wide variety of surgical tools
and surgeries must be captured for robust training. Moreover, ongoing annotator
training is needed to stay up to date with surgical instrument innovation. In
robotic-assisted surgery, however, potentially informative data like timestamps
of instrument installation and removal can be programmatically harvested. The
ability to rely on tool installation data alone would significantly reduce the
workload to train robust tool-tracking models. With this motivation in mind we
invited the surgical data science community to participate in the challenge,
SurgToolLoc 2022. The goal was to leverage tool presence data as weak labels
for machine learning models trained to detect tools and localize them in video
frames with bounding boxes. We present the results of this challenge along with
many of the team's efforts. We conclude by discussing these results in the
broader context of machine learning and surgical data science. The training
data used for this challenge consisting of 24,695 video clips with tool
presence labels is also being released publicly and can be accessed at
https://console.cloud.google.com/storage/browser/isi-surgtoolloc-2022.
Related papers
- SAR-RARP50: Segmentation of surgical instrumentation and Action
Recognition on Robot-Assisted Radical Prostatectomy Challenge [72.97934765570069]
We release the first multimodal, publicly available, in-vivo, dataset for surgical action recognition and semantic instrumentation segmentation, containing 50 suturing video segments of Robotic Assisted Radical Prostatectomy (RARP)
The aim of the challenge is to enable researchers to leverage the scale of the provided dataset and develop robust and highly accurate single-task action recognition and tool segmentation approaches in the surgical domain.
A total of 12 teams participated in the challenge, contributing 7 action recognition methods, 9 instrument segmentation techniques, and 4 multitask approaches that integrated both action recognition and instrument segmentation.
arXiv Detail & Related papers (2023-12-31T13:32:18Z) - Computer Vision for Increased Operative Efficiency via Identification of Instruments in the Neurosurgical Operating Room: A Proof-of-Concept Study [33.41411356073856]
Computer vision (CV) is a field of artificial intelligence that enables machines to interpret and understand images and videos.
CV has the potential to be of assistance in the operating room (OR) to track surgical instruments.
We built a CV algorithm for identifying surgical instruments in the neurosurgical operating room.
arXiv Detail & Related papers (2023-12-03T19:01:50Z) - Demonstration-Guided Reinforcement Learning with Efficient Exploration
for Task Automation of Surgical Robot [54.80144694888735]
We introduce Demonstration-guided EXploration (DEX), an efficient reinforcement learning algorithm.
Our method estimates expert-like behaviors with higher values to facilitate productive interactions.
Experiments on $10$ surgical manipulation tasks from SurRoL, a comprehensive surgical simulation platform, demonstrate significant improvements.
arXiv Detail & Related papers (2023-02-20T05:38:54Z) - Objective Surgical Skills Assessment and Tool Localization: Results from
the MICCAI 2021 SimSurgSkill Challenge [11.007322707874184]
SimSurgSkill 2021 (hosted as a sub-challenge of EndoVis at MICCAI 2021) aimed to promote and foster work in this endeavor.
Competitors were tasked with localizing instruments and predicting surgical skill.
Using this publicly available dataset and results as a springboard, future work may enable more efficient training of surgeons with advances in surgical data science.
arXiv Detail & Related papers (2022-12-08T18:14:52Z) - Rethinking Surgical Instrument Segmentation: A Background Image Can Be
All You Need [18.830738606514736]
Data scarcity and imbalance have heavily affected the model accuracy and limited the design and deployment of deep learning-based surgical applications.
We propose a one-to-many data generation solution that gets rid of the complicated and expensive process of data collection and annotation from robotic surgery.
Our empirical analysis suggests that without the high cost of data collection and annotation, we can achieve decent surgical instrument segmentation performance.
arXiv Detail & Related papers (2022-06-23T16:22:56Z) - Relational Graph Learning on Visual and Kinematics Embeddings for
Accurate Gesture Recognition in Robotic Surgery [84.73764603474413]
We propose a novel online approach of multi-modal graph network (i.e., MRG-Net) to dynamically integrate visual and kinematics information.
The effectiveness of our method is demonstrated with state-of-the-art results on the public JIGSAWS dataset.
arXiv Detail & Related papers (2020-11-03T11:00:10Z) - Visual Imitation Made Easy [102.36509665008732]
We present an alternate interface for imitation that simplifies the data collection process while allowing for easy transfer to robots.
We use commercially available reacher-grabber assistive tools both as a data collection device and as the robot's end-effector.
We experimentally evaluate on two challenging tasks: non-prehensile pushing and prehensile stacking, with 1000 diverse demonstrations for each task.
arXiv Detail & Related papers (2020-08-11T17:58:50Z) - Synthetic and Real Inputs for Tool Segmentation in Robotic Surgery [10.562627972607892]
We show that it may be possible to use robot kinematic data coupled with laparoscopic images to alleviate the labelling problem.
We propose a new deep learning based model for parallel processing of both laparoscopic and simulation images.
arXiv Detail & Related papers (2020-07-17T16:33:33Z) - Recurrent and Spiking Modeling of Sparse Surgical Kinematics [0.8458020117487898]
A growing number of studies have used machine learning to analyze video and kinematic data captured from surgical robots.
In this study, we explore the possibility of using only kinematic data to predict surgeons of similar skill levels.
We report that it is possible to identify surgical fellows receiving near perfect scores in the simulation exercises based on their motion characteristics alone.
arXiv Detail & Related papers (2020-05-12T15:41:45Z) - LRTD: Long-Range Temporal Dependency based Active Learning for Surgical
Workflow Recognition [67.86810761677403]
We propose a novel active learning method for cost-effective surgical video analysis.
Specifically, we propose a non-local recurrent convolutional network (NL-RCNet), which introduces non-local block to capture the long-range temporal dependency.
We validate our approach on a large surgical video dataset (Cholec80) by performing surgical workflow recognition task.
arXiv Detail & Related papers (2020-04-21T09:21:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.