See More, Change Less: Anatomy-Aware Diffusion for Contrast Enhancement
- URL: http://arxiv.org/abs/2512.07251v1
- Date: Mon, 08 Dec 2025 07:48:45 GMT
- Title: See More, Change Less: Anatomy-Aware Diffusion for Contrast Enhancement
- Authors: Junqi Liu, Zejun Wu, Pedro R. A. S. Bassi, Xinze Zhou, Wenxuan Li, Ibrahim E. Hamamci, Sezgin Er, Tianyu Lin, Yi Luo, Szymon Płotka, Bjoern Menze, Daguang Xu, Kai Ding, Kang Wang, Yang Yang, Yucheng Tang, Alan L. Yuille, Zongwei Zhou,
- Abstract summary: SMILE is an anatomy-aware diffusion model that learns how organs are shaped and how they take up contrast.<n>It enhances only clinically relevant regions while leaving all other areas unchanged.<n>It also improves cancer detection from non-contrast CT, raising the F1 score by up to 10 percent.
- Score: 54.01053990883076
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Image enhancement improves visual quality and helps reveal details that are hard to see in the original image. In medical imaging, it can support clinical decision-making, but current models often over-edit. This can distort organs, create false findings, and miss small tumors because these models do not understand anatomy or contrast dynamics. We propose SMILE, an anatomy-aware diffusion model that learns how organs are shaped and how they take up contrast. It enhances only clinically relevant regions while leaving all other areas unchanged. SMILE introduces three key ideas: (1) structure-aware supervision that follows true organ boundaries and contrast patterns; (2) registration-free learning that works directly with unaligned multi-phase CT scans; (3) unified inference that provides fast and consistent enhancement across all contrast phases. Across six external datasets, SMILE outperforms existing methods in image quality (14.2% higher SSIM, 20.6% higher PSNR, 50% better FID) and in clinical usefulness by producing anatomically accurate and diagnostically meaningful images. SMILE also improves cancer detection from non-contrast CT, raising the F1 score by up to 10 percent.
Related papers
- Structure-constrained Language-informed Diffusion Model for Unpaired Low-dose Computed Tomography Angiography Reconstruction [72.80209358480424]
overdose of iodinated contrast media (ICM) can cause kidney damage and life-threatening allergic reactions.<n>Deep learning methods can generate CT images of normal-dose ICM from low-dose ICM, reducing the required dose.<n>We propose a Structure-constrained Language-informed Diffusion Model (SLDM) that integrates structural synergy and spatial intelligence.
arXiv Detail & Related papers (2026-01-28T06:54:06Z) - Dual Interaction Network with Cross-Image Attention for Medical Image Segmentation [4.804685266333253]
We propose a dual interactive fusion module (DIFM) that exploits mutual complementary information from the original and enhanced images.<n>DIFM employs cross-attention bidirectionally to simultaneously attend to corresponding spatial information across different images.<n>In addition, we introduce a multi-scale boundary loss based on gradient extraction to improve segmentation accuracy at object boundaries.
arXiv Detail & Related papers (2025-09-07T07:25:59Z) - PLUS: Plug-and-Play Enhanced Liver Lesion Diagnosis Model on Non-Contrast CT Scans [21.598723671897766]
PLUS is a plug-and-play framework that enhances FLL analysis on NCCT images for arbitrary 3D segmentation models.<n>Results: PLUS improved lesion-level F1 score by 5.66%, the malignant patient-level F1 score by 6.26%, and the benign patient-level F1 score by 4.03%.
arXiv Detail & Related papers (2025-07-05T03:03:13Z) - Optimizing Breast Cancer Detection in Mammograms: A Comprehensive Study of Transfer Learning, Resolution Reduction, and Multi-View Classification [0.0]
Mammography, an X-ray-based imaging technique, remains central to the early detection of breast cancer.<n>Recent advances in artificial intelligence have enabled increasingly sophisticated computer-aided diagnostic methods.<n>Despite this progress, several critical questions remain unanswered.
arXiv Detail & Related papers (2025-03-25T11:51:21Z) - Diffusion Models for Counterfactual Generation and Anomaly Detection in Brain Images [39.94162291765236]
We present a weakly supervised method to generate a healthy version of a diseased image and then use it to obtain a pixel-wise anomaly map.
We employ a diffusion model trained on healthy samples and combine Denoising Diffusion Probabilistic Model (DDPM) and Denoising Implicit Model (DDIM) at each step of the sampling process.
arXiv Detail & Related papers (2023-08-03T21:56:50Z) - Learning Better Contrastive View from Radiologist's Gaze [45.55702035003462]
We propose a novel augmentation method, i.e., FocusContrast, to learn from radiologists' gaze in diagnosis and generate contrastive views for medical images.
Specifically, we track the gaze movement of radiologists and model their visual attention when reading to diagnose X-ray images.
As a plug-and-play module, FocusContrast consistently improves state-of-the-art contrastive learning methods of SimCLR, MoCo, and BYOL by 4.07.0% in classification accuracy on a knee X-ray dataset.
arXiv Detail & Related papers (2023-05-15T17:34:49Z) - Mine yOur owN Anatomy: Revisiting Medical Image Segmentation with Extremely Limited Labels [54.58539616385138]
We introduce a novel semi-supervised 2D medical image segmentation framework termed Mine yOur owN Anatomy (MONA)
First, prior work argues that every pixel equally matters to the model training; we observe empirically that this alone is unlikely to define meaningful anatomical features.
Second, we construct a set of objectives that encourage the model to be capable of decomposing medical images into a collection of anatomical features.
arXiv Detail & Related papers (2022-09-27T15:50:31Z) - Anatomy-Aware Contrastive Representation Learning for Fetal Ultrasound [17.91546880972773]
We propose to improve visual representations of medical images via anatomy-aware contrastive learning (AWCL)
AWCL incorporates anatomy information to augment the positive/negative pair sampling in a contrastive learning manner.
Experiments on a large-scale fetal ultrasound dataset demonstrate that our approach is effective for learning representations that transfer well to three clinical downstream tasks.
arXiv Detail & Related papers (2022-08-22T22:49:26Z) - Harmonizing Pathological and Normal Pixels for Pseudo-healthy Synthesis [68.5287824124996]
We present a new type of discriminator, the segmentor, to accurately locate the lesions and improve the visual quality of pseudo-healthy images.
We apply the generated images into medical image enhancement and utilize the enhanced results to cope with the low contrast problem.
Comprehensive experiments on the T2 modality of BraTS demonstrate that the proposed method substantially outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T08:41:17Z) - Improving Tuberculosis (TB) Prediction using Synthetically Generated
Computed Tomography (CT) Images [0.17499351967216337]
Pulmonary infections can often be best imaged and evaluated through computed tomography (CT) scans.
X-ray, a different type of imaging procedure, is inexpensive, often available at the bedside and more widely available, but offers a simpler, two dimensional image.
We show that by relying on a model that learns to generate CT images from X-rays synthetically, we can improve the automatic disease classification accuracy.
arXiv Detail & Related papers (2021-09-23T16:35:15Z) - A Multi-Stage Attentive Transfer Learning Framework for Improving
COVID-19 Diagnosis [49.3704402041314]
We propose a multi-stage attentive transfer learning framework for improving COVID-19 diagnosis.
Our proposed framework consists of three stages to train accurate diagnosis models through learning knowledge from multiple source tasks and data of different domains.
Importantly, we propose a novel self-supervised learning method to learn multi-scale representations for lung CT images.
arXiv Detail & Related papers (2021-01-14T01:39:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.