Intelligent Masking: Deep Q-Learning for Context Encoding in Medical
Image Analysis
- URL: http://arxiv.org/abs/2203.13865v1
- Date: Fri, 25 Mar 2022 19:05:06 GMT
- Title: Intelligent Masking: Deep Q-Learning for Context Encoding in Medical
Image Analysis
- Authors: Mojtaba Bahrami, Mahsa Ghorbani, Nassir Navab
- Abstract summary: We develop a novel self-supervised approach that occludes targeted regions to improve the pre-training procedure.
We show that training the agent against the prediction model can significantly improve the semantic features extracted for downstream classification tasks.
- Score: 48.02011627390706
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The need for a large amount of labeled data in the supervised setting has led
recent studies to utilize self-supervised learning to pre-train deep neural
networks using unlabeled data. Many self-supervised training strategies have
been investigated especially for medical datasets to leverage the information
available in the much fewer unlabeled data. One of the fundamental strategies
in image-based self-supervision is context prediction. In this approach, a
model is trained to reconstruct the contents of an arbitrary missing region of
an image based on its surroundings. However, the existing methods adopt a
random and blind masking approach by focusing uniformly on all regions of the
images. This approach results in a lot of unnecessary network updates that
cause the model to forget the rich extracted features. In this work, we develop
a novel self-supervised approach that occludes targeted regions to improve the
pre-training procedure. To this end, we propose a reinforcement learning-based
agent which learns to intelligently mask input images through deep Q-learning.
We show that training the agent against the prediction model can significantly
improve the semantic features extracted for downstream classification tasks. We
perform our experiments on two public datasets for diagnosing breast cancer in
the ultrasound images and detecting lower-grade glioma with MR images. In our
experiments, we show that our novel masking strategy advances the learned
features according to the performance on the classification task in terms of
accuracy, macro F1, and AUROC.
Related papers
- Local-to-Global Self-Supervised Representation Learning for Diabetic Retinopathy Grading [0.0]
This research aims to present a novel hybrid learning model using self-supervised learning and knowledge distillation.
In our algorithm, for the first time among all self-supervised learning and knowledge distillation models, the test dataset is 50% larger than the training dataset.
Compared to a similar state-of-the-art model, our results achieved higher accuracy and more effective representation spaces.
arXiv Detail & Related papers (2024-10-01T15:19:16Z) - Nucleus-aware Self-supervised Pretraining Using Unpaired Image-to-image
Translation for Histopathology Images [3.8391355786589805]
We propose a novel nucleus-aware self-supervised pretraining framework for histopathology images.
The framework aims to capture the nuclear morphology and distribution information through unpaired image-to-image translation.
The experiments on 7 datasets show that the proposed pretraining method outperforms supervised ones on Kather classification, multiple instance learning, and 5 dense-prediction tasks.
arXiv Detail & Related papers (2023-09-14T02:31:18Z) - Free-ATM: Exploring Unsupervised Learning on Diffusion-Generated Images
with Free Attention Masks [64.67735676127208]
Text-to-image diffusion models have shown great potential for benefiting image recognition.
Although promising, there has been inadequate exploration dedicated to unsupervised learning on diffusion-generated images.
We introduce customized solutions by fully exploiting the aforementioned free attention masks.
arXiv Detail & Related papers (2023-08-13T10:07:46Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - Improving Masked Autoencoders by Learning Where to Mask [65.89510231743692]
Masked image modeling is a promising self-supervised learning method for visual data.
We present AutoMAE, a framework that uses Gumbel-Softmax to interlink an adversarially-trained mask generator and a mask-guided image modeling process.
In our experiments, AutoMAE is shown to provide effective pretraining models on standard self-supervised benchmarks and downstream tasks.
arXiv Detail & Related papers (2023-03-12T05:28:55Z) - Self-Supervised-RCNN for Medical Image Segmentation with Limited Data
Annotation [0.16490701092527607]
We propose an alternative deep learning training strategy based on self-supervised pretraining on unlabeled MRI scans.
Our pretraining approach first, randomly applies different distortions to random areas of unlabeled images and then predicts the type of distortions and loss of information.
The effectiveness of the proposed method for segmentation tasks in different pre-training and fine-tuning scenarios is evaluated.
arXiv Detail & Related papers (2022-07-17T13:28:52Z) - Deep reinforced active learning for multi-class image classification [0.0]
High accuracy medical image classification can be limited by the costs of acquiring more data as well as the time and expertise needed to label existing images.
We apply active learning to medical image classification, a method which aims to maximise model performance on a minimal subset from a larger pool of data.
arXiv Detail & Related papers (2022-06-20T09:30:55Z) - Continual Learning for Blind Image Quality Assessment [80.55119990128419]
Blind image quality assessment (BIQA) models fail to continually adapt to subpopulation shift.
Recent work suggests training BIQA methods on the combination of all available human-rated IQA datasets.
We formulate continual learning for BIQA, where a model learns continually from a stream of IQA datasets.
arXiv Detail & Related papers (2021-02-19T03:07:01Z) - Self supervised contrastive learning for digital histopathology [0.0]
We use a contrastive self-supervised learning method called SimCLR that achieved state-of-the-art results on natural-scene images.
We find that combining multiple multi-organ datasets with different types of staining and resolution properties improves the quality of the learned features.
Linear classifiers trained on top of the learned features show that networks pretrained on digital histopathology datasets perform better than ImageNet pretrained networks.
arXiv Detail & Related papers (2020-11-27T19:18:45Z) - Improved Slice-wise Tumour Detection in Brain MRIs by Computing
Dissimilarities between Latent Representations [68.8204255655161]
Anomaly detection for Magnetic Resonance Images (MRIs) can be solved with unsupervised methods.
We have proposed a slice-wise semi-supervised method for tumour detection based on the computation of a dissimilarity function in the latent space of a Variational AutoEncoder.
We show that by training the models on higher resolution images and by improving the quality of the reconstructions, we obtain results which are comparable with different baselines.
arXiv Detail & Related papers (2020-07-24T14:02:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.