An Automated Machine Learning Framework for Surgical Suturing Action Detection under Class Imbalance
- URL: http://arxiv.org/abs/2502.06407v1
- Date: Mon, 10 Feb 2025 12:47:36 GMT
- Title: An Automated Machine Learning Framework for Surgical Suturing Action Detection under Class Imbalance
- Authors: Baobing Zhang, Paul Sullivan, Benjie Tang, Ghulam Nabi, Mustafa Suphi Erden,
- Abstract summary: Real-time detection of surgical actions with interpretable outputs is crucial for automated and real-time instructional feedback and skill development.
This paper presents a rapid deployment approach utilizing automated machine learning methods, based on surgical action data collected from both experienced and trainee surgeons.
- Score: 1.2043621020930133
- License:
- Abstract: In laparoscopy surgical training and evaluation, real-time detection of surgical actions with interpretable outputs is crucial for automated and real-time instructional feedback and skill development. Such capability would enable development of machine guided training systems. This paper presents a rapid deployment approach utilizing automated machine learning methods, based on surgical action data collected from both experienced and trainee surgeons. The proposed approach effectively tackles the challenge of highly imbalanced class distributions, ensuring robust predictions across varying skill levels of surgeons. Additionally, our method partially incorporates model transparency, addressing the reliability requirements in medical applications. Compared to deep learning approaches, traditional machine learning models not only facilitate efficient rapid deployment but also offer significant advantages in interpretability. Through experiments, this study demonstrates the potential of this approach to provide quick, reliable and effective real-time detection in surgical training environments
Related papers
- Multi-Modal Self-Supervised Learning for Surgical Feedback Effectiveness Assessment [66.6041949490137]
We propose a method that integrates information from transcribed verbal feedback and corresponding surgical video to predict feedback effectiveness.
Our findings show that both transcribed feedback and surgical video are individually predictive of trainee behavior changes.
Our results demonstrate the potential of multi-modal learning to advance the automated assessment of surgical feedback.
arXiv Detail & Related papers (2024-11-17T00:13:00Z) - Learning Hand State Estimation for a Light Exoskeleton [50.05509088121445]
We propose a machine learning-based estimator of the hand state for rehabilitation purposes, using light exoskeletons.
We build a supervised approach using information from the muscular activity of the forearm and the motion of the exoskeleton to reconstruct the hand's opening degree and compliance level.
Our approach is validated with a real light exoskeleton.
arXiv Detail & Related papers (2024-11-14T09:12:38Z) - Video-based Surgical Skill Assessment using Tree-based Gaussian Process
Classifier [2.3964255330849356]
This paper presents a novel pipeline for automated surgical skill assessment using video data.
The pipeline incorporates a representation flow convolutional neural network and a novel tree-based Gaussian process classifier.
The proposed method has the potential to facilitate skill improvement among surgery fellows and enhance patient safety.
arXiv Detail & Related papers (2023-12-15T21:06:22Z) - Surgical tool classification and localization: results and methods from
the MICCAI 2022 SurgToolLoc challenge [69.91670788430162]
We present the results of the SurgLoc 2022 challenge.
The goal was to leverage tool presence data as weak labels for machine learning models trained to detect tools.
We conclude by discussing these results in the broader context of machine learning and surgical data science.
arXiv Detail & Related papers (2023-05-11T21:44:39Z) - Uncertainty-aware Self-supervised Learning for Cross-domain Technical
Skill Assessment in Robot-assisted Surgery [14.145726158070522]
We propose a novel approach for skill assessment by transferring domain knowledge from labeled kinematic data to unlabeled data.
Our method offers a significant advantage over other existing works as it does not require manual labeling or prior knowledge of the surgical training task for robot-assisted surgery.
arXiv Detail & Related papers (2023-04-28T01:52:18Z) - Demonstration-Guided Reinforcement Learning with Efficient Exploration
for Task Automation of Surgical Robot [54.80144694888735]
We introduce Demonstration-guided EXploration (DEX), an efficient reinforcement learning algorithm.
Our method estimates expert-like behaviors with higher values to facilitate productive interactions.
Experiments on $10$ surgical manipulation tasks from SurRoL, a comprehensive surgical simulation platform, demonstrate significant improvements.
arXiv Detail & Related papers (2023-02-20T05:38:54Z) - Video-based Surgical Skills Assessment using Long term Tool Tracking [0.3324986723090368]
We introduce a motion-based approach to automatically assess surgical skills from surgical case video feed.
The proposed pipeline first tracks surgical tools reliably to create motion trajectories.
We compare transformer-based skill assessment with traditional machine learning approaches using the proposed and state-of-the-art tracking.
arXiv Detail & Related papers (2022-07-05T18:15:28Z) - Real-time Informative Surgical Skill Assessment with Gaussian Process
Learning [12.019641896240245]
This work presents a novel Gaussian Process Learning-based automatic objective surgical skill assessment method for ESSBSs.
The proposed method projects the instrument movements into the endoscope coordinate to reduce the data dimensionality.
The experimental results show that the proposed method reaches 100% prediction precision for complete surgical procedures and 90% precision for real-time prediction assessment.
arXiv Detail & Related papers (2021-12-05T15:35:40Z) - Adversarial Training is Not Ready for Robot Learning [55.493354071227174]
Adversarial training is an effective method to train deep learning models that are resilient to norm-bounded perturbations.
We show theoretically and experimentally that neural controllers obtained via adversarial training are subjected to three types of defects.
Our results suggest that adversarial training is not yet ready for robot learning.
arXiv Detail & Related papers (2021-03-15T07:51:31Z) - Automatic Gesture Recognition in Robot-assisted Surgery with
Reinforcement Learning and Tree Search [63.07088785532908]
We propose a framework based on reinforcement learning and tree search for joint surgical gesture segmentation and classification.
Our framework consistently outperforms the existing methods on the suturing task of JIGSAWS dataset in terms of accuracy, edit score and F1 score.
arXiv Detail & Related papers (2020-02-20T13:12:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.