Contrastive Learning with Continuous Proxy Meta-Data for 3D MRI
Classification
- URL: http://arxiv.org/abs/2106.08808v1
- Date: Wed, 16 Jun 2021 14:17:04 GMT
- Title: Contrastive Learning with Continuous Proxy Meta-Data for 3D MRI
Classification
- Authors: Benoit Dufumier, Pietro Gori, Julie Victor, Antoine Grigis, Michel
Wessa, Paolo Brambilla, Pauline Favre, Mircea Polosan, Colm McDonald, Camille
Marie Piguet, Edouard Duchesnay
- Abstract summary: We propose to leverage continuous proxy metadata, in the contrastive learning framework, by introducing a new loss called y-Aware InfoNCE loss.
A 3D CNN model pre-trained on $104$ multi-site healthy brain MRI scans can extract relevant features for three classification tasks.
When fine-tuned, it also outperforms 3D CNN trained from scratch on these tasks, as well as state-of-the-art self-supervised methods.
- Score: 1.714108629548376
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional supervised learning with deep neural networks requires a
tremendous amount of labelled data to converge to a good solution. For 3D
medical images, it is often impractical to build a large homogeneous annotated
dataset for a specific pathology. Self-supervised methods offer a new way to
learn a representation of the images in an unsupervised manner with a neural
network. In particular, contrastive learning has shown great promises by
(almost) matching the performance of fully-supervised CNN on vision tasks.
Nonetheless, this method does not take advantage of available meta-data, such
as participant's age, viewed as prior knowledge. Here, we propose to leverage
continuous proxy metadata, in the contrastive learning framework, by
introducing a new loss called y-Aware InfoNCE loss. Specifically, we improve
the positive sampling during pre-training by adding more positive examples with
similar proxy meta-data with the anchor, assuming they share similar
discriminative semantic features.With our method, a 3D CNN model pre-trained on
$10^4$ multi-site healthy brain MRI scans can extract relevant features for
three classification tasks: schizophrenia, bipolar diagnosis and Alzheimer's
detection. When fine-tuned, it also outperforms 3D CNN trained from scratch on
these tasks, as well as state-of-the-art self-supervised methods. Our code is
made publicly available here.
Related papers
- Enhancing Weakly Supervised 3D Medical Image Segmentation through
Probabilistic-aware Learning [52.249748801637196]
3D medical image segmentation is a challenging task with crucial implications for disease diagnosis and treatment planning.
Recent advances in deep learning have significantly enhanced fully supervised medical image segmentation.
We propose a novel probabilistic-aware weakly supervised learning pipeline, specifically designed for 3D medical imaging.
arXiv Detail & Related papers (2024-03-05T00:46:53Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - ALSO: Automotive Lidar Self-supervision by Occupancy estimation [70.70557577874155]
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds.
The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled.
The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information.
arXiv Detail & Related papers (2022-12-12T13:10:19Z) - Intelligent Masking: Deep Q-Learning for Context Encoding in Medical
Image Analysis [48.02011627390706]
We develop a novel self-supervised approach that occludes targeted regions to improve the pre-training procedure.
We show that training the agent against the prediction model can significantly improve the semantic features extracted for downstream classification tasks.
arXiv Detail & Related papers (2022-03-25T19:05:06Z) - Benchmarking CNN on 3D Anatomical Brain MRI: Architectures, Data
Augmentation and Deep Ensemble Learning [2.1446056201053185]
We propose an extensive benchmark of recent state-of-the-art (SOTA) 3D CNN, evaluating also the benefits of data augmentation and deep ensemble learning.
Experiments were conducted on a large multi-site 3D brain anatomical MRI data-set comprising N=10k scans on 3 challenging tasks: age prediction, sex classification, and schizophrenia diagnosis.
We found that all models provide significantly better predictions with VBM images than quasi-raw data.
DenseNet and tiny-DenseNet, a lighter version that we proposed, provide a good compromise in terms of performance in all data regime
arXiv Detail & Related papers (2021-06-02T13:00:35Z) - Convolution-Free Medical Image Segmentation using Transformers [8.130670465411239]
We show that a different method, based entirely on self-attention between neighboring image patches, can achieve competitive or better results.
We show that the proposed model can achieve segmentation accuracies that are better than the state of the art CNNs on three datasets.
arXiv Detail & Related papers (2021-02-26T18:49:13Z) - Uniformizing Techniques to Process CT scans with 3D CNNs for
Tuberculosis Prediction [5.270882613122642]
A common approach to medical image analysis on volumetric data uses deep 2D convolutional neural networks (CNNs)
dealing with the individual slices independently in 2D CNNs deliberately discards the depth information which results in poor performance for the intended task.
We evaluate a set of volume uniformizing methods to address the aforementioned issues.
We report 73% area under curve (AUC) and binary classification accuracy (ACC) of 67.5% on the test set beating all methods which leveraged only image information.
arXiv Detail & Related papers (2020-07-26T21:53:47Z) - Embedding Task Knowledge into 3D Neural Networks via Self-supervised
Learning [21.902313057142905]
Self-supervised learning (SSL) is a potential solution for deficient annotated data.
We propose a novel SSL approach for 3D medical image classification, namely Task-related Contrastive Prediction Coding ( TCPC)
TCPC embeds task knowledge into training 3D neural networks.
arXiv Detail & Related papers (2020-06-10T12:37:39Z) - 3D Self-Supervised Methods for Medical Imaging [7.65168530693281]
We propose 3D versions for five different self-supervised methods, in the form of proxy tasks.
Our methods facilitate neural network feature learning from unlabeled 3D images, aiming to reduce the required cost for expert annotation.
The developed algorithms are 3D Contrastive Predictive Coding, 3D Rotation prediction, 3D Jigsaw puzzles, Relative 3D patch location, and 3D Exemplar networks.
arXiv Detail & Related papers (2020-06-06T09:56:58Z) - 3D medical image segmentation with labeled and unlabeled data using
autoencoders at the example of liver segmentation in CT images [58.720142291102135]
This work investigates the potential of autoencoder-extracted features to improve segmentation with a convolutional neural network.
A convolutional autoencoder was used to extract features from unlabeled data and a multi-scale, fully convolutional CNN was used to perform the target task of 3D liver segmentation in CT images.
arXiv Detail & Related papers (2020-03-17T20:20:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.