DeepSet SimCLR: Self-supervised deep sets for improved pathology
representation learning
- URL: http://arxiv.org/abs/2402.15598v1
- Date: Fri, 23 Feb 2024 20:37:59 GMT
- Title: DeepSet SimCLR: Self-supervised deep sets for improved pathology
representation learning
- Authors: David Torpey and Richard Klein
- Abstract summary: We aim to improve standard 2D SSL algorithms by modelling the inherent 3D nature of these datasets implicitly.
We propose two variants that build upon a strong baseline model and show that both of these variants often outperform the baseline in a variety of downstream tasks.
- Score: 4.40560654491339
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Often, applications of self-supervised learning to 3D medical data opt to use
3D variants of successful 2D network architectures. Although promising
approaches, they are significantly more computationally demanding to train, and
thus reduce the widespread applicability of these methods away from those with
modest computational resources. Thus, in this paper, we aim to improve standard
2D SSL algorithms by modelling the inherent 3D nature of these datasets
implicitly. We propose two variants that build upon a strong baseline model and
show that both of these variants often outperform the baseline in a variety of
downstream tasks. Importantly, in contrast to previous works in both 2D and 3D
approaches for 3D medical data, both of our proposals introduce negligible
additional overhead over the baseline, improving the democratisation of these
approaches for medical applications.
Related papers
- Cross-D Conv: Cross-Dimensional Transferable Knowledge Base via Fourier Shifting Operation [3.69758875412828]
Cross-D Conv operation bridges the dimensional gap by learning the phase shifting in the Fourier domain.
Our method enables seamless weight transfer between 2D and 3D convolution operations, effectively facilitating cross-dimensional learning.
arXiv Detail & Related papers (2024-11-02T13:03:44Z) - Pretrained Deep 2.5D Models for Efficient Predictive Modeling from
Retinal OCT [7.8641166297532035]
3D deep learning models play a crucial role in building powerful predictive models of disease progression.
In this paper, we explore 2.5D architectures based on a combination of convolutional neural networks (CNNs), long short-term memory (LSTM), and Transformers.
We demonstrate the effectiveness of architectures and associated pretraining on a task of predicting progression to wet age-related macular degeneration (AMD) within a six-month period.
arXiv Detail & Related papers (2023-07-25T23:46:48Z) - Interpretable 2D Vision Models for 3D Medical Images [47.75089895500738]
This study proposes a simple approach of adapting 2D networks with an intermediate feature representation for processing 3D images.
We show on all 3D MedMNIST datasets as benchmark and two real-world datasets consisting of several hundred high-resolution CT or MRI scans that our approach performs on par with existing methods.
arXiv Detail & Related papers (2023-07-13T08:27:09Z) - Improving 3D Imaging with Pre-Trained Perpendicular 2D Diffusion Models [52.529394863331326]
We propose a novel approach using two perpendicular pre-trained 2D diffusion models to solve the 3D inverse problem.
Our method is highly effective for 3D medical image reconstruction tasks, including MRI Z-axis super-resolution, compressed sensing MRI, and sparse-view CT.
arXiv Detail & Related papers (2023-03-15T08:28:06Z) - Deep Generative Models on 3D Representations: A Survey [81.73385191402419]
Generative models aim to learn the distribution of observed data by generating new instances.
Recently, researchers started to shift focus from 2D to 3D space.
representing 3D data poses significantly greater challenges.
arXiv Detail & Related papers (2022-10-27T17:59:50Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - Improved Brain Age Estimation with Slice-based Set Networks [18.272915375351914]
We propose a new architecture for BrainAGE prediction.
The proposed architecture works by encoding each 2D slice in an MRI with a deep 2D-CNN model.
Next, it combines the information from these 2D-slice encodings using set networks or permutation invariant layers.
Experiments on the BrainAGE prediction problem, using the UK Biobank dataset, showed that the model with the permutation invariant layers trains faster and provides better predictions compared to other state-of-the-art approaches.
arXiv Detail & Related papers (2021-02-08T18:54:15Z) - Secrets of 3D Implicit Object Shape Reconstruction in the Wild [92.5554695397653]
Reconstructing high-fidelity 3D objects from sparse, partial observation is crucial for various applications in computer vision, robotics, and graphics.
Recent neural implicit modeling methods show promising results on synthetic or dense datasets.
But, they perform poorly on real-world data that is sparse and noisy.
This paper analyzes the root cause of such deficient performance of a popular neural implicit model.
arXiv Detail & Related papers (2021-01-18T03:24:48Z) - Volumetric Medical Image Segmentation: A 3D Deep Coarse-to-fine
Framework and Its Adversarial Examples [74.92488215859991]
We propose a novel 3D-based coarse-to-fine framework to efficiently tackle these challenges.
The proposed 3D-based framework outperforms their 2D counterparts by a large margin since it can leverage the rich spatial information along all three axes.
We conduct experiments on three datasets, the NIH pancreas dataset, the JHMI pancreas dataset and the JHMI pathological cyst dataset.
arXiv Detail & Related papers (2020-10-29T15:39:19Z) - Exemplar Fine-Tuning for 3D Human Model Fitting Towards In-the-Wild 3D
Human Pose Estimation [107.07047303858664]
Large-scale human datasets with 3D ground-truth annotations are difficult to obtain in the wild.
We address this problem by augmenting existing 2D datasets with high-quality 3D pose fits.
The resulting annotations are sufficient to train from scratch 3D pose regressor networks that outperform the current state-of-the-art on in-the-wild benchmarks.
arXiv Detail & Related papers (2020-04-07T20:21:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.