Prediction of Geometric Transformation on Cardiac MRI via Convolutional
Neural Network
- URL: http://arxiv.org/abs/2211.06641v1
- Date: Sat, 12 Nov 2022 11:29:14 GMT
- Title: Prediction of Geometric Transformation on Cardiac MRI via Convolutional
Neural Network
- Authors: Xin Gao
- Abstract summary: We propose to learn features in medical images by training ConvNets to recognize the geometric transformation applied to images.
We present a simple self-supervised task that can easily predict the geometric transformation.
- Score: 13.01021780124613
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the field of medical image, deep convolutional neural networks(ConvNets)
have achieved great success in the classification, segmentation, and
registration tasks thanks to their unparalleled capacity to learn image
features. However, these tasks often require large amounts of manually
annotated data and are labor-intensive. Therefore, it is of significant
importance for us to study unsupervised semantic feature learning tasks. In our
work, we propose to learn features in medical images by training ConvNets to
recognize the geometric transformation applied to images and present a simple
self-supervised task that can easily predict the geometric transformation. We
precisely define a set of geometric transformations in mathematical terms and
generalize this model to 3D, taking into account the distinction between
spatial and time dimensions. We evaluated our self-supervised method on CMR
images of different modalities (bSSFP, T2, LGE) and achieved accuracies of
96.4%, 97.5%, and 96.4%, respectively. The code and models of our paper will be
published on: https://github.com/gaoxin492/Geometric_Transformation_CMR
Related papers
- Large Spatial Model: End-to-end Unposed Images to Semantic 3D [79.94479633598102]
Large Spatial Model (LSM) processes unposed RGB images directly into semantic radiance fields.
LSM simultaneously estimates geometry, appearance, and semantics in a single feed-forward operation.
It can generate versatile label maps by interacting with language at novel viewpoints.
arXiv Detail & Related papers (2024-10-24T17:54:42Z) - Cross-domain and Cross-dimension Learning for Image-to-Graph
Transformers [50.576354045312115]
Direct image-to-graph transformation is a challenging task that solves object detection and relationship prediction in a single model.
We introduce a set of methods enabling cross-domain and cross-dimension transfer learning for image-to-graph transformers.
We demonstrate our method's utility in cross-domain and cross-dimension experiments, where we pretrain our models on 2D satellite images before applying them to vastly different target domains in 2D and 3D.
arXiv Detail & Related papers (2024-03-11T10:48:56Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - TransformNet: Self-supervised representation learning through predicting
geometric transformations [0.8098097078441623]
We describe the unsupervised semantic feature learning approach for recognition of the geometric transformation applied to the input data.
The basic concept of our approach is that if someone is unaware of the objects in the images, he/she would not be able to quantitatively predict the geometric transformation that was applied to them.
arXiv Detail & Related papers (2022-02-08T22:41:01Z) - Leveraging Unsupervised Image Registration for Discovery of Landmark
Shape Descriptor [5.40076482533193]
This paper proposes a self-supervised deep learning approach for discovering landmarks from images that can directly be used as a shape descriptor for subsequent analysis.
We use landmark-driven image registration as the primary task to force the neural network to discover landmarks that register the images well.
The proposed method circumvents segmentation and preprocessing and directly produces a usable shape descriptor using just 2D or 3D images.
arXiv Detail & Related papers (2021-11-13T01:02:10Z) - Automatic size and pose homogenization with spatial transformer network
to improve and accelerate pediatric segmentation [51.916106055115755]
We propose a new CNN architecture that is pose and scale invariant thanks to the use of Spatial Transformer Network (STN)
Our architecture is composed of three sequential modules that are estimated together during training.
We test the proposed method in kidney and renal tumor segmentation on abdominal pediatric CT scanners.
arXiv Detail & Related papers (2021-07-06T14:50:03Z) - Deep learning based geometric registration for medical images: How
accurate can we get without visual features? [5.05806585671215]
Deep learning is driving the development of new approaches for image registration.
In this work we take a look at an exactly opposite approach by investigating a deep learning framework for registration based solely on geometric features and optimisation.
Our experimental validation is conducted on complex key-point graphs of inner lung structures, strongly outperforming dense encoder-decoder networks and other point set registration methods.
arXiv Detail & Related papers (2021-03-01T10:15:47Z) - TSGCNet: Discriminative Geometric Feature Learning with Two-Stream
GraphConvolutional Network for 3D Dental Model Segmentation [141.2690520327948]
We propose a two-stream graph convolutional network (TSGCNet) to learn multi-view information from different geometric attributes.
We evaluate our proposed TSGCNet on a real-patient dataset of dental models acquired by 3D intraoral scanners.
arXiv Detail & Related papers (2020-12-26T08:02:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.