Unleashing the Power of Depth and Pose Estimation Neural Networks by
Designing Compatible Endoscopic Images
- URL: http://arxiv.org/abs/2309.07390v1
- Date: Thu, 14 Sep 2023 02:19:38 GMT
- Title: Unleashing the Power of Depth and Pose Estimation Neural Networks by
Designing Compatible Endoscopic Images
- Authors: Junyang Wu, Yun Gu
- Abstract summary: We conduct a detail analysis of the properties of endoscopic images and improve the compatibility of images and neural networks.
First, we introcude the Mask Image Modelling (MIM) module, which inputs partial image information instead of complete image information.
Second, we propose a lightweight neural network to enhance the endoscopic images, to explicitly improve the compatibility between images and neural networks.
- Score: 12.412060445862842
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning models have witnessed depth and pose estimation framework on
unannotated datasets as a effective pathway to succeed in endoscopic
navigation. Most current techniques are dedicated to developing more advanced
neural networks to improve the accuracy. However, existing methods ignore the
special properties of endoscopic images, resulting in an inability to fully
unleash the power of neural networks. In this study, we conduct a detail
analysis of the properties of endoscopic images and improve the compatibility
of images and neural networks, to unleash the power of current neural networks.
First, we introcude the Mask Image Modelling (MIM) module, which inputs partial
image information instead of complete image information, allowing the network
to recover global information from partial pixel information. This enhances the
network' s ability to perceive global information and alleviates the phenomenon
of local overfitting in convolutional neural networks due to local artifacts.
Second, we propose a lightweight neural network to enhance the endoscopic
images, to explicitly improve the compatibility between images and neural
networks. Extensive experiments are conducted on the three public datasets and
one inhouse dataset, and the proposed modules improve baselines by a large
margin. Furthermore, the enhanced images we proposed, which have higher network
compatibility, can serve as an effective data augmentation method and they are
able to extract more stable feature points in traditional feature point
matching tasks and achieve outstanding performance.
Related papers
- Research on Deep Learning Model of Feature Extraction Based on Convolutional Neural Network [0.32985979395737786]
AlexNet and InceptionV3 were selected to obtain better image recognition results.
The prediction accuracy, specificity, and sensitivity of the trained AlexNet model increased by 4.25 percentage points.
The graphics processing usage has decreased by 51% compared to the InceptionV3 mode.
arXiv Detail & Related papers (2024-06-13T06:00:28Z) - DmADs-Net: Dense multiscale attention and depth-supervised network for medical image segmentation [10.85494240952418]
We have created the Multiscale Attention and Depth-Supervised Network (DmADs-Net)
We use ResNet for feature extraction at different depths and create a Multi-scale Convolutional Feature Attention Block.
The Local Feature Attention Block is created to enable enhanced local feature attention for high-level semantic information.
In the feature fusion phase, a Feature Refinement and Fusion Block is created to enhance the fusion of different semantic information.
arXiv Detail & Related papers (2024-05-01T12:15:58Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Retinotopic Mapping Enhances the Robustness of Convolutional Neural Networks [0.0]
This study investigates whether retinotopic mapping, a critical component of foveated vision, can enhance image categorization and localization performance.
Renotopic mapping was integrated into the inputs of standard off-the-shelf convolutional neural networks (CNNs)
Surprisingly, the retinotopically mapped network achieved comparable performance in classification.
arXiv Detail & Related papers (2024-02-23T18:15:37Z) - Learning Multimodal Volumetric Features for Large-Scale Neuron Tracing [72.45257414889478]
We aim to reduce human workload by predicting connectivity between over-segmented neuron pieces.
We first construct a dataset, named FlyTracing, that contains millions of pairwise connections of segments expanding the whole fly brain.
We propose a novel connectivity-aware contrastive learning method to generate dense volumetric EM image embedding.
arXiv Detail & Related papers (2024-01-05T19:45:12Z) - Connecting the Dots: Graph Neural Network Powered Ensemble and
Classification of Medical Images [0.0]
Deep learning for medical imaging is limited due to the requirement for large amounts of training data.
We employ the Image Foresting Transform to optimally segment images into superpixels.
These superpixels are subsequently transformed into graph-structured data, enabling the proficient extraction of features and modeling of relationships.
arXiv Detail & Related papers (2023-11-13T13:20:54Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Domain Generalization for Medical Imaging Classification with
Linear-Dependency Regularization [59.5104563755095]
We introduce a simple but effective approach to improve the generalization capability of deep neural networks in the field of medical imaging classification.
Motivated by the observation that the domain variability of the medical images is to some extent compact, we propose to learn a representative feature space through variational encoding.
arXiv Detail & Related papers (2020-09-27T12:30:30Z) - A new approach to descriptors generation for image retrieval by
analyzing activations of deep neural network layers [43.77224853200986]
We consider the problem of descriptors construction for the task of content-based image retrieval using deep neural networks.
It is known that the total number of neurons in the convolutional part of the network is large and the majority of them have little influence on the final classification decision.
We propose a novel algorithm that allows us to extract the most significant neuron activations and utilize this information to construct effective descriptors.
arXiv Detail & Related papers (2020-07-13T18:53:10Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.