Uncertainty-aware Self-supervised Learning for Cross-domain Technical
Skill Assessment in Robot-assisted Surgery
- URL: http://arxiv.org/abs/2304.14589v1
- Date: Fri, 28 Apr 2023 01:52:18 GMT
- Title: Uncertainty-aware Self-supervised Learning for Cross-domain Technical
Skill Assessment in Robot-assisted Surgery
- Authors: Ziheng Wang, Andrea Mariani, Arianna Menciassi, Elena De Momi, Ann
Majewicz Fey
- Abstract summary: We propose a novel approach for skill assessment by transferring domain knowledge from labeled kinematic data to unlabeled data.
Our method offers a significant advantage over other existing works as it does not require manual labeling or prior knowledge of the surgical training task for robot-assisted surgery.
- Score: 14.145726158070522
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Objective technical skill assessment is crucial for effective training of new
surgeons in robot-assisted surgery. With advancements in surgical training
programs in both physical and virtual environments, it is imperative to develop
generalizable methods for automatically assessing skills. In this paper, we
propose a novel approach for skill assessment by transferring domain knowledge
from labeled kinematic data to unlabeled data. Our approach leverages labeled
data from common surgical training tasks such as Suturing, Needle Passing, and
Knot Tying to jointly train a model with both labeled and unlabeled data.
Pseudo labels are generated for the unlabeled data through an iterative manner
that incorporates uncertainty estimation to ensure accurate labeling. We
evaluate our method on a virtual reality simulated training task (Ring
Transfer) using data from the da Vinci Research Kit (dVRK). The results show
that trainees with robotic assistance have significantly higher expert
probability compared to these without any assistance, p < 0.05, which aligns
with previous studies showing the benefits of robotic assistance in improving
training proficiency. Our method offers a significant advantage over other
existing works as it does not require manual labeling or prior knowledge of the
surgical training task for robot-assisted surgery.
Related papers
- Multi-Modal Self-Supervised Learning for Surgical Feedback Effectiveness Assessment [66.6041949490137]
We propose a method that integrates information from transcribed verbal feedback and corresponding surgical video to predict feedback effectiveness.
Our findings show that both transcribed feedback and surgical video are individually predictive of trainee behavior changes.
Our results demonstrate the potential of multi-modal learning to advance the automated assessment of surgical feedback.
arXiv Detail & Related papers (2024-11-17T00:13:00Z) - Robust Surgical Tools Detection in Endoscopic Videos with Noisy Data [2.566694420723775]
We propose a systematic methodology for developing robust models for surgical tool detection using noisy data.
Our methodology introduces two key innovations: (1) an intelligent active learning strategy for minimal dataset identification and label correction by human experts; and (2) an assembling strategy for a student-teacher model-based self-training framework.
The proposed methodology achieves an average F1-score of 85.88% for the ensemble model-based self-training with class weights, and 80.88% without class weights for noisy labels.
arXiv Detail & Related papers (2023-07-03T08:12:56Z) - Demonstration-Guided Reinforcement Learning with Efficient Exploration
for Task Automation of Surgical Robot [54.80144694888735]
We introduce Demonstration-guided EXploration (DEX), an efficient reinforcement learning algorithm.
Our method estimates expert-like behaviors with higher values to facilitate productive interactions.
Experiments on $10$ surgical manipulation tasks from SurRoL, a comprehensive surgical simulation platform, demonstrate significant improvements.
arXiv Detail & Related papers (2023-02-20T05:38:54Z) - Video-based Surgical Skills Assessment using Long term Tool Tracking [0.3324986723090368]
We introduce a motion-based approach to automatically assess surgical skills from surgical case video feed.
The proposed pipeline first tracks surgical tools reliably to create motion trajectories.
We compare transformer-based skill assessment with traditional machine learning approaches using the proposed and state-of-the-art tracking.
arXiv Detail & Related papers (2022-07-05T18:15:28Z) - Revisiting the Adversarial Robustness-Accuracy Tradeoff in Robot
Learning [121.9708998627352]
Recent work has shown that, in practical robot learning applications, the effects of adversarial training do not pose a fair trade-off.
This work revisits the robustness-accuracy trade-off in robot learning by analyzing if recent advances in robust training methods and theory can make adversarial training suitable for real-world robot applications.
arXiv Detail & Related papers (2022-04-15T08:12:15Z) - Federated Cycling (FedCy): Semi-supervised Federated Learning of
Surgical Phases [57.90226879210227]
FedCy is a semi-supervised learning (FSSL) method that combines FL and self-supervised learning to exploit a decentralized dataset of both labeled and unlabeled videos.
We demonstrate significant performance gains over state-of-the-art FSSL methods on the task of automatic recognition of surgical phases.
arXiv Detail & Related papers (2022-03-14T17:44:53Z) - Real-time Informative Surgical Skill Assessment with Gaussian Process
Learning [12.019641896240245]
This work presents a novel Gaussian Process Learning-based automatic objective surgical skill assessment method for ESSBSs.
The proposed method projects the instrument movements into the endoscope coordinate to reduce the data dimensionality.
The experimental results show that the proposed method reaches 100% prediction precision for complete surgical procedures and 90% precision for real-time prediction assessment.
arXiv Detail & Related papers (2021-12-05T15:35:40Z) - Domain Adaptive Robotic Gesture Recognition with Unsupervised
Kinematic-Visual Data Alignment [60.31418655784291]
We propose a novel unsupervised domain adaptation framework which can simultaneously transfer multi-modality knowledge, i.e., both kinematic and visual data, from simulator to real robot.
It remedies the domain gap with enhanced transferable features by using temporal cues in videos, and inherent correlations in multi-modal towards recognizing gesture.
Results show that our approach recovers the performance with great improvement gains, up to 12.91% in ACC and 20.16% in F1score without using any annotations in real robot.
arXiv Detail & Related papers (2021-03-06T09:10:03Z) - Surgical Skill Assessment on In-Vivo Clinical Data via the Clearness of
Operating Field [18.643159726513133]
Surgical skill assessment is studied in this paper on a real clinical dataset.
The clearness of operating field (COF) is identified as a good proxy for overall surgical skills.
An objective and automated framework is proposed to predict surgical skills through the proxy of COF.
In experiments, the proposed method achieves 0.55 Spearman's correlation with the ground truth of overall technical skill.
arXiv Detail & Related papers (2020-08-27T07:12:16Z) - Scalable Multi-Task Imitation Learning with Autonomous Improvement [159.9406205002599]
We build an imitation learning system that can continuously improve through autonomous data collection.
We leverage the robot's own trials as demonstrations for tasks other than the one that the robot actually attempted.
In contrast to prior imitation learning approaches, our method can autonomously collect data with sparse supervision for continuous improvement.
arXiv Detail & Related papers (2020-02-25T18:56:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.