Bounded Future MS-TCN++ for surgical gesture recognition
- URL: http://arxiv.org/abs/2209.14647v1
- Date: Thu, 29 Sep 2022 09:09:54 GMT
- Title: Bounded Future MS-TCN++ for surgical gesture recognition
- Authors: Adam Goldbraikh, Netanell Avisdris, Carla M. Pugh, Shlomi Laufer
- Abstract summary: We learn the performance-delay trade-off and design an MS-TCN++-based algorithm that can utilize this trade-off.
The naive approach is to reduce the MS-TCN++ depth, as a result, the receptive field is reduced, and also the number of required future frames is also reduced.
This way, we have flexibility in the network design and as a result, we achieve significantly better performance than in the naive approach.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In recent times there is a growing development of video based applications
for surgical purposes. Part of these applications can work offline after the
end of the procedure, other applications must react immediately. However, there
are cases where the response should be done during the procedure but some delay
is acceptable. In the literature, the online-offline performance gap is known.
Our goal in this study was to learn the performance-delay trade-off and design
an MS-TCN++-based algorithm that can utilize this trade-off. To this aim, we
used our open surgery simulation data-set containing 96 videos of 24
participants that perform a suturing task on a variable tissue simulator. In
this study, we used video data captured from the side view. The Networks were
trained to identify the performed surgical gestures. The naive approach is to
reduce the MS-TCN++ depth, as a result, the receptive field is reduced, and
also the number of required future frames is also reduced. We showed that this
method is sub-optimal, mainly in the small delay cases. The second method was
to limit the accessible future in each temporal convolution. This way, we have
flexibility in the network design and as a result, we achieve significantly
better performance than in the naive approach.
Related papers
- Efficient Deformable Tissue Reconstruction via Orthogonal Neural Plane [58.871015937204255]
We introduce Fast Orthogonal Plane (plane) for the reconstruction of deformable tissues.
We conceptualize surgical procedures as 4D volumes, and break them down into static and dynamic fields comprised of neural planes.
This factorization iscretizes four-dimensional space, leading to a decreased memory usage and faster optimization.
arXiv Detail & Related papers (2023-12-23T13:27:50Z) - Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Neural Network Pruning for Real-time Polyp Segmentation [8.08470060885395]
We show an application of neural network pruning in polyp segmentation.
We compute the importance score of convolutional filters and remove the filters having the least scores, which to some value of pruning does not degrade the performance.
arXiv Detail & Related papers (2023-06-22T21:03:50Z) - Fast Neural Scene Flow [36.29234109363439]
A coordinate neural network estimates scene flow at runtime, without any training.
In this paper, we demonstrate that scene flow is different -- with the dominant computational bottleneck stemming from the loss function itself.
Our fast neural scene flow (FNSF) approach reports for the first time real-time performance comparable to learning methods.
arXiv Detail & Related papers (2023-04-18T16:37:18Z) - AEGNN: Asynchronous Event-based Graph Neural Networks [54.528926463775946]
Event-based Graph Neural Networks generalize standard GNNs to process events as "evolving"-temporal graphs.
AEGNNs are easily trained on synchronous inputs and can be converted to efficient, "asynchronous" networks at test time.
arXiv Detail & Related papers (2022-03-31T16:21:12Z) - Learning Bayesian Sparse Networks with Full Experience Replay for
Continual Learning [54.7584721943286]
Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered.
Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal.
We propose to only activate and select sparse neurons for learning current and past tasks at any stage.
arXiv Detail & Related papers (2022-02-21T13:25:03Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - SurgeonAssist-Net: Towards Context-Aware Head-Mounted Display-Based
Augmented Reality for Surgical Guidance [18.060445966264727]
SurgeonAssist-Net is a framework making action-and-workflow-driven virtual assistance accessible to commercially available optical see-through head-mounted displays (OST-HMDs)
Our implementation competes with state-of-the-art approaches in prediction accuracy for automated task recognition.
It is capable of near real-time performance on the Microsoft HoloLens 2 OST-HMD.
arXiv Detail & Related papers (2021-07-13T21:12:34Z) - LRTD: Long-Range Temporal Dependency based Active Learning for Surgical
Workflow Recognition [67.86810761677403]
We propose a novel active learning method for cost-effective surgical video analysis.
Specifically, we propose a non-local recurrent convolutional network (NL-RCNet), which introduces non-local block to capture the long-range temporal dependency.
We validate our approach on a large surgical video dataset (Cholec80) by performing surgical workflow recognition task.
arXiv Detail & Related papers (2020-04-21T09:21:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.