View it like a radiologist: Shifted windows for deep learning
augmentation of CT images
- URL: http://arxiv.org/abs/2311.14990v1
- Date: Sat, 25 Nov 2023 10:28:08 GMT
- Title: View it like a radiologist: Shifted windows for deep learning
augmentation of CT images
- Authors: Eirik A. {\O}stmo, Kristoffer K. Wickstr{\o}m, Keyur Radiya, Michael
C. Kampffmeyer, Robert Jenssen
- Abstract summary: We propose a novel preprocessing and intensity augmentation scheme inspired by how radiologists leverage multiple viewing windows when evaluating CT images.
Our proposed method, window shifting, randomly places the viewing windows around the region of interest during training.
This approach improves liver lesion segmentation performance and robustness on images with poorly timed contrast agent.
- Score: 11.902593645631034
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning has the potential to revolutionize medical practice by
automating and performing important tasks like detecting and delineating the
size and locations of cancers in medical images. However, most deep learning
models rely on augmentation techniques that treat medical images as natural
images. For contrast-enhanced Computed Tomography (CT) images in particular,
the signals producing the voxel intensities have physical meaning, which is
lost during preprocessing and augmentation when treating such images as natural
images. To address this, we propose a novel preprocessing and intensity
augmentation scheme inspired by how radiologists leverage multiple viewing
windows when evaluating CT images. Our proposed method, window shifting,
randomly places the viewing windows around the region of interest during
training. This approach improves liver lesion segmentation performance and
robustness on images with poorly timed contrast agent. Our method outperforms
classical intensity augmentations as well as the intensity augmentation
pipeline of the popular nn-UNet on multiple datasets.
Related papers
- Autoregressive Sequence Modeling for 3D Medical Image Representation [48.706230961589924]
We introduce a pioneering method for learning 3D medical image representations through an autoregressive sequence pre-training framework.
Our approach various 3D medical images based on spatial, contrast, and semantic correlations, treating them as interconnected visual tokens within a token sequence.
arXiv Detail & Related papers (2024-09-13T10:19:10Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - LOTUS: Learning to Optimize Task-based US representations [39.81131738128329]
Anatomical segmentation of organs in ultrasound images is essential to many clinical applications.
Existing deep neural networks require a large amount of labeled data for training in order to achieve clinically acceptable performance.
In this paper, we propose a novel approach for learning to optimize task-based ultra-sound image representations.
arXiv Detail & Related papers (2023-07-29T16:29:39Z) - Learning Better Contrastive View from Radiologist's Gaze [45.55702035003462]
We propose a novel augmentation method, i.e., FocusContrast, to learn from radiologists' gaze in diagnosis and generate contrastive views for medical images.
Specifically, we track the gaze movement of radiologists and model their visual attention when reading to diagnose X-ray images.
As a plug-and-play module, FocusContrast consistently improves state-of-the-art contrastive learning methods of SimCLR, MoCo, and BYOL by 4.07.0% in classification accuracy on a knee X-ray dataset.
arXiv Detail & Related papers (2023-05-15T17:34:49Z) - Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - Exploring Image Augmentations for Siamese Representation Learning with
Chest X-Rays [0.8808021343665321]
We train and evaluate Siamese Networks for abnormality detection on chest X-Rays.
We identify a set of augmentations that yield robust representations that generalize well to both out-of-distribution data and diseases.
arXiv Detail & Related papers (2023-01-30T03:42:02Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z) - Pathological Retinal Region Segmentation From OCT Images Using Geometric
Relation Based Augmentation [84.7571086566595]
We propose improvements over previous GAN-based medical image synthesis methods by jointly encoding the intrinsic relationship of geometry and shape.
The proposed method outperforms state-of-the-art segmentation methods on the public RETOUCH dataset having images captured from different acquisition procedures.
arXiv Detail & Related papers (2020-03-31T11:50:43Z) - Lesion Conditional Image Generation for Improved Segmentation of
Intracranial Hemorrhage from CT Images [0.0]
We present a lesion conditional Generative Adversarial Network LcGAN to generate synthetic Computed Tomography (CT) images for data augmentation.
A lesion conditional image (segmented mask) is an input to both the generator and the discriminator of the LcGAN during training.
We quantify the quality of the images by using a fully convolutional network (FCN) score and blurriness.
arXiv Detail & Related papers (2020-03-30T23:32:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.