DA-VSR: Domain Adaptable Volumetric Super-Resolution For Medical Images
- URL: http://arxiv.org/abs/2210.05117v1
- Date: Tue, 11 Oct 2022 03:16:35 GMT
- Title: DA-VSR: Domain Adaptable Volumetric Super-Resolution For Medical Images
- Authors: Cheng Peng, S. Kevin Zhou, and Rama Chellappa
- Abstract summary: We present a novel algorithm called domain adaptable super-resolution (DA-VSR) to better bridge the domain inconsistency gap.
DA-VSR uses a unified feature extraction backbone and a series of network heads to improve image quality over different planes.
We demonstrate that DA-VSR significantly improves super-resolution quality across numerous datasets of different domains.
- Score: 69.63915773870758
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical image super-resolution (SR) is an active research area that has many
potential applications, including reducing scan time, bettering visual
understanding, increasing robustness in downstream tasks, etc. However,
applying deep-learning-based SR approaches for clinical applications often
encounters issues of domain inconsistency, as the test data may be acquired by
different machines or on different organs. In this work, we present a novel
algorithm called domain adaptable volumetric super-resolution (DA-VSR) to
better bridge the domain inconsistency gap. DA-VSR uses a unified feature
extraction backbone and a series of network heads to improve image quality over
different planes. Furthermore, DA-VSR leverages the in-plane and through-plane
resolution differences on the test data to achieve a self-learned domain
adaptation. As such, DA-VSR combines the advantages of a strong feature
generator learned through supervised training and the ability to tune to the
idiosyncrasies of the test volumes through unsupervised learning. Through
experiments, we demonstrate that DA-VSR significantly improves super-resolution
quality across numerous datasets of different domains, thereby taking a further
step toward real clinical applications.
Related papers
- MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language Models [49.765466293296186]
Recent progress in Medical Large Vision-Language Models (Med-LVLMs) has opened up new possibilities for interactive diagnostic tools.
Med-LVLMs often suffer from factual hallucination, which can lead to incorrect diagnoses.
We propose a versatile multimodal RAG system, MMed-RAG, designed to enhance the factuality of Med-LVLMs.
arXiv Detail & Related papers (2024-10-16T23:03:27Z) - Curriculum-Based Augmented Fourier Domain Adaptation for Robust Medical
Image Segmentation [18.830738606514736]
This work proposes the Curriculum-based Augmented Fourier Domain Adaptation (Curri-AFDA) for robust medical image segmentation.
In particular, our curriculum learning strategy is based on the causal relationship of a model under different levels of data shift.
Experiments on two segmentation tasks of Retina and Nuclei collected from multiple sites and scanners suggest that our proposed method yields superior adaptation and generalization performance.
arXiv Detail & Related papers (2023-06-06T08:56:58Z) - Multi-scale Feature Alignment for Continual Learning of Unlabeled
Domains [3.9498537297431167]
generative feature-driven image replay in conjunction with a dual-purpose discriminator enables the generation of images with realistic features for replay.
We present detailed ablation experiments studying our proposed method components and demonstrate a possible use-case of our continual UDA method for an unsupervised patch-based segmentation task.
arXiv Detail & Related papers (2023-02-02T18:19:01Z) - RPLHR-CT Dataset and Transformer Baseline for Volumetric
Super-Resolution from CT Scans [12.066026343488453]
coarse resolution may lead to difficulties in medical diagnosis by either physicians or computer-aided diagnosis algorithms.
Deep learning-based volumetric super-resolution (SR) methods are feasible ways to improve resolution.
This paper builds the first public real-paired dataset RPLHR-CT as a benchmark for volumetric SR.
Considering the inherent shortcoming of CNN, we also propose a transformer volumetric super-resolution network (TVSRN) based on attention mechanisms.
arXiv Detail & Related papers (2022-06-13T15:35:59Z) - Studying Robustness of Semantic Segmentation under Domain Shift in
cardiac MRI [0.8858288982748155]
We study challenges and opportunities of domain transfer across images from multiple clinical centres and scanner vendors.
In this work, we build upon a fixed U-Net architecture configured by the nnU-net framework to investigate various data augmentation techniques and batch normalization layers.
arXiv Detail & Related papers (2020-11-15T17:50:23Z) - Deep Representational Similarity Learning for analyzing neural
signatures in task-based fMRI dataset [81.02949933048332]
This paper develops Deep Representational Similarity Learning (DRSL), a deep extension of Representational Similarity Analysis (RSA)
DRSL is appropriate for analyzing similarities between various cognitive tasks in fMRI datasets with a large number of subjects.
arXiv Detail & Related papers (2020-09-28T18:30:14Z) - Domain-Adversarial Learning for Multi-Centre, Multi-Vendor, and
Multi-Disease Cardiac MR Image Segmentation [3.4551186283197883]
Domain-adversarial learning is used to train a domain-invariant 2D U-Net using labelled and unlabelled data.
This approach is evaluated on both seen and unseen domains from the M&Ms challenge dataset.
arXiv Detail & Related papers (2020-08-26T19:40:55Z) - MuCAN: Multi-Correspondence Aggregation Network for Video
Super-Resolution [63.02785017714131]
Video super-resolution (VSR) aims to utilize multiple low-resolution frames to generate a high-resolution prediction for each frame.
Inter- and intra-frames are the key sources for exploiting temporal and spatial information.
We build an effective multi-correspondence aggregation network (MuCAN) for VSR.
arXiv Detail & Related papers (2020-07-23T05:41:27Z) - Cross-domain Face Presentation Attack Detection via Multi-domain
Disentangled Representation Learning [109.42987031347582]
Face presentation attack detection (PAD) has been an urgent problem to be solved in the face recognition systems.
We propose an efficient disentangled representation learning for cross-domain face PAD.
Our approach consists of disentangled representation learning (DR-Net) and multi-domain learning (MD-Net)
arXiv Detail & Related papers (2020-04-04T15:45:14Z) - MS-Net: Multi-Site Network for Improving Prostate Segmentation with
Heterogeneous MRI Data [75.73881040581767]
We propose a novel multi-site network (MS-Net) for improving prostate segmentation by learning robust representations.
Our MS-Net improves the performance across all datasets consistently, and outperforms state-of-the-art methods for multi-site learning.
arXiv Detail & Related papers (2020-02-09T14:11:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.