Adapting to Unseen Vendor Domains for MRI Lesion Segmentation
- URL: http://arxiv.org/abs/2108.06434v1
- Date: Sat, 14 Aug 2021 01:09:43 GMT
- Title: Adapting to Unseen Vendor Domains for MRI Lesion Segmentation
- Authors: Brandon Mac, Alan R. Moody, April Khademi
- Abstract summary: We investigate an unsupervised image translation model to augment MR images from a source dataset to a target dataset.
We consider three configurations of augmentation between datasets consisting of translation between images, between scanner vendors, and from labels to images.
It was found that the segmentation models trained on synthetic data from labels to images configuration yielded the closest performance to the segmentation model trained directly on the target dataset.
- Score: 0.08156494881838945
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One of the key limitations in machine learning models is poor performance on
data that is out of the domain of the training distribution. This is especially
true for image analysis in magnetic resonance (MR) imaging, as variations in
hardware and software create non-standard intensities, contrasts, and noise
distributions across scanners. Recently, image translation models have been
proposed to augment data across domains to create synthetic data points. In
this paper, we investigate the application an unsupervised image translation
model to augment MR images from a source dataset to a target dataset.
Specifically, we want to evaluate how well these models can create synthetic
data points representative of the target dataset through image translation, and
to see if a segmentation model trained these synthetic data points would
approach the performance of a model trained directly on the target dataset. We
consider three configurations of augmentation between datasets consisting of
translation between images, between scanner vendors, and from labels to images.
It was found that the segmentation models trained on synthetic data from labels
to images configuration yielded the closest performance to the segmentation
model trained directly on the target dataset. The Dice coeffcient score per
each target vendor (GE, Siemens, Philips) for training on synthetic data was
0.63, 0.64, and 0.58, compared to training directly on target dataset was 0.65,
0.72, and 0.61.
Related papers
- DataDream: Few-shot Guided Dataset Generation [90.09164461462365]
We propose a framework for synthesizing classification datasets that more faithfully represents the real data distribution.
DataDream fine-tunes LoRA weights for the image generation model on the few real images before generating the training data using the adapted model.
We then fine-tune LoRA weights for CLIP using the synthetic data to improve downstream image classification over previous approaches on a large variety of datasets.
arXiv Detail & Related papers (2024-07-15T17:10:31Z) - Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - DataDAM: Efficient Dataset Distillation with Attention Matching [15.300968899043498]
Researchers have long tried to minimize training costs in deep learning by maintaining strong generalization across diverse datasets.
Emerging research on dataset aims to reduce training costs by creating a small synthetic set that contains the information of a larger real dataset.
However, the synthetic data generated by previous methods are not guaranteed to distribute and discriminate as well as the original training data.
arXiv Detail & Related papers (2023-09-29T19:07:48Z) - DatasetDM: Synthesizing Data with Perception Annotations Using Diffusion
Models [61.906934570771256]
We present a generic dataset generation model that can produce diverse synthetic images and perception annotations.
Our method builds upon the pre-trained diffusion model and extends text-guided image synthesis to perception data generation.
We show that the rich latent code of the diffusion model can be effectively decoded as accurate perception annotations using a decoder module.
arXiv Detail & Related papers (2023-08-11T14:38:11Z) - Generating Reliable Pixel-Level Labels for Source Free Domain Adaptation [13.913151437401472]
ReGEN comprises an image-to-image translation network and a segmentation network.
Our workflow generates target-like images using the noisy predictions from the original target domain images.
arXiv Detail & Related papers (2023-07-03T09:44:13Z) - Zero-shot Composed Text-Image Retrieval [72.43790281036584]
We consider the problem of composed image retrieval (CIR)
It aims to train a model that can fuse multi-modal information, e.g., text and images, to accurately retrieve images that match the query, extending the user's expression ability.
arXiv Detail & Related papers (2023-06-12T17:56:01Z) - Unsupervised Domain Adaptation with Histogram-gated Image Translation
for Delayered IC Image Analysis [2.720699926154399]
Histogram-gated Image Translation (HGIT) is an unsupervised domain adaptation framework which transforms images from a given source dataset to the domain of a target dataset.
Our method achieves the best performance compared to the reported domain adaptation techniques, and is also reasonably close to the fully supervised benchmark.
arXiv Detail & Related papers (2022-09-27T15:53:22Z) - Prefix Conditioning Unifies Language and Label Supervision [84.11127588805138]
We show that dataset biases negatively affect pre-training by reducing the generalizability of learned representations.
In experiments, we show that this simple technique improves the performance in zero-shot image recognition accuracy and robustness to the image-level distribution shift.
arXiv Detail & Related papers (2022-06-02T16:12:26Z) - MSeg: A Composite Dataset for Multi-domain Semantic Segmentation [100.17755160696939]
We present MSeg, a composite dataset that unifies semantic segmentation datasets from different domains.
We reconcile the generalization and bring the pixel-level annotations into alignment by relabeling more than 220,000 object masks in more than 80,000 images.
A model trained on MSeg ranks first on the WildDash-v1 leaderboard for robust semantic segmentation, with no exposure to WildDash data during training.
arXiv Detail & Related papers (2021-12-27T16:16:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.