SuPer Deep: A Surgical Perception Framework for Robotic Tissue
Manipulation using Deep Learning for Feature Extraction
- URL: http://arxiv.org/abs/2003.03472v3
- Date: Thu, 25 Mar 2021 05:41:43 GMT
- Title: SuPer Deep: A Surgical Perception Framework for Robotic Tissue
Manipulation using Deep Learning for Feature Extraction
- Authors: Jingpei Lu, Ambareesh Jayakumari, Florian Richter, Yang Li, Michael C.
Yip
- Abstract summary: We exploit deep learning methods for surgical perception.
We integrated deep neural networks, capable of efficient feature extraction, into the tissue reconstruction and instrument pose estimation processes.
Our framework achieves state-of-the-art tracking performance in a surgical environment by utilizing deep learning for feature extraction.
- Score: 25.865648975312407
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Robotic automation in surgery requires precise tracking of surgical tools and
mapping of deformable tissue. Previous works on surgical perception frameworks
require significant effort in developing features for surgical tool and tissue
tracking. In this work, we overcome the challenge by exploiting deep learning
methods for surgical perception. We integrated deep neural networks, capable of
efficient feature extraction, into the tissue reconstruction and instrument
pose estimation processes. By leveraging transfer learning, the deep learning
based approach requires minimal training data and reduced feature engineering
efforts to fully perceive a surgical scene. The framework was tested on three
publicly available datasets, which use the da Vinci Surgical System, for
comprehensive analysis. Experimental results show that our framework achieves
state-of-the-art tracking performance in a surgical environment by utilizing
deep learning for feature extraction.
Related papers
- Creating a Digital Twin of Spinal Surgery: A Proof of Concept [68.37190859183663]
Surgery digitalization is the process of creating a virtual replica of real-world surgery.
We present a proof of concept (PoC) for surgery digitalization that is applied to an ex-vivo spinal surgery.
We employ five RGB-D cameras for dynamic 3D reconstruction of the surgeon, a high-end camera for 3D reconstruction of the anatomy, an infrared stereo camera for surgical instrument tracking, and a laser scanner for 3D reconstruction of the operating room and data fusion.
arXiv Detail & Related papers (2024-03-25T13:09:40Z) - Autonomous Soft Tissue Retraction Using Demonstration-Guided
Reinforcement Learning [6.80186731352488]
Existing surgical task learning mainly pertains to rigid body interactions.
The advancement towards more sophisticated surgical robots necessitates the manipulation of soft bodies.
This work lays the foundation for future research into the development and refinement of surgical robots capable of managing both rigid and soft tissue interactions.
arXiv Detail & Related papers (2023-09-02T06:13:58Z) - SAMSNeRF: Segment Anything Model (SAM) Guides Dynamic Surgical Scene
Reconstruction by Neural Radiance Field (NeRF) [4.740415113160021]
We propose a novel approach called SAMSNeRF that combines Segment Anything Model (SAM) and Neural Radiance Field (NeRF) techniques.
Our experimental results on public endoscopy surgical videos demonstrate that our approach successfully reconstructs high-fidelity dynamic surgical scenes.
arXiv Detail & Related papers (2023-08-22T20:31:00Z) - Dynamic Interactive Relation Capturing via Scene Graph Learning for
Robotic Surgical Report Generation [14.711668177329244]
For robot-assisted surgery, an accurate surgical report reflects clinical operations during surgery and helps document entry tasks, post-operative analysis and follow-up treatment.
It is a challenging task due to many complex and diverse interactions between instruments and tissues in the surgical scene.
This paper presents a neural network to boost surgical report generation by explicitly exploring the interactive relation between tissues and surgical instruments.
arXiv Detail & Related papers (2023-06-05T07:34:41Z) - Surgical tool classification and localization: results and methods from
the MICCAI 2022 SurgToolLoc challenge [69.91670788430162]
We present the results of the SurgLoc 2022 challenge.
The goal was to leverage tool presence data as weak labels for machine learning models trained to detect tools.
We conclude by discussing these results in the broader context of machine learning and surgical data science.
arXiv Detail & Related papers (2023-05-11T21:44:39Z) - Demonstration-Guided Reinforcement Learning with Efficient Exploration
for Task Automation of Surgical Robot [54.80144694888735]
We introduce Demonstration-guided EXploration (DEX), an efficient reinforcement learning algorithm.
Our method estimates expert-like behaviors with higher values to facilitate productive interactions.
Experiments on $10$ surgical manipulation tasks from SurRoL, a comprehensive surgical simulation platform, demonstrate significant improvements.
arXiv Detail & Related papers (2023-02-20T05:38:54Z) - Robotic Navigation Autonomy for Subretinal Injection via Intelligent
Real-Time Virtual iOCT Volume Slicing [88.99939660183881]
We propose a framework for autonomous robotic navigation for subretinal injection.
Our method consists of an instrument pose estimation method, an online registration between the robotic and the i OCT system, and trajectory planning tailored for navigation to an injection target.
Our experiments on ex-vivo porcine eyes demonstrate the precision and repeatability of the method.
arXiv Detail & Related papers (2023-01-17T21:41:21Z) - CholecTriplet2021: A benchmark challenge for surgical action triplet
recognition [66.51610049869393]
This paper presents CholecTriplet 2021: an endoscopic vision challenge organized at MICCAI 2021 for the recognition of surgical action triplets in laparoscopic videos.
We present the challenge setup and assessment of the state-of-the-art deep learning methods proposed by the participants during the challenge.
A total of 4 baseline methods and 19 new deep learning algorithms are presented to recognize surgical action triplets directly from surgical videos, achieving mean average precision (mAP) ranging from 4.2% to 38.1%.
arXiv Detail & Related papers (2022-04-10T18:51:55Z) - E-DSSR: Efficient Dynamic Surgical Scene Reconstruction with
Transformer-based Stereoscopic Depth Perception [15.927060244702686]
We present an efficient reconstruction pipeline for highly dynamic surgical scenes that runs at 28 fps.
Specifically, we design a transformer-based stereoscopic depth perception for efficient depth estimation.
We evaluate the proposed pipeline on two datasets, the public Hamlyn Centre Endoscopic Video dataset and our in-house DaVinci robotic surgery dataset.
arXiv Detail & Related papers (2021-07-01T05:57:41Z) - Synthetic and Real Inputs for Tool Segmentation in Robotic Surgery [10.562627972607892]
We show that it may be possible to use robot kinematic data coupled with laparoscopic images to alleviate the labelling problem.
We propose a new deep learning based model for parallel processing of both laparoscopic and simulation images.
arXiv Detail & Related papers (2020-07-17T16:33:33Z) - Automatic Gesture Recognition in Robot-assisted Surgery with
Reinforcement Learning and Tree Search [63.07088785532908]
We propose a framework based on reinforcement learning and tree search for joint surgical gesture segmentation and classification.
Our framework consistently outperforms the existing methods on the suturing task of JIGSAWS dataset in terms of accuracy, edit score and F1 score.
arXiv Detail & Related papers (2020-02-20T13:12:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.