Feature-Conditioned Cascaded Video Diffusion Models for Precise
Echocardiogram Synthesis
- URL: http://arxiv.org/abs/2303.12644v3
- Date: Wed, 21 Feb 2024 17:56:30 GMT
- Title: Feature-Conditioned Cascaded Video Diffusion Models for Precise
Echocardiogram Synthesis
- Authors: Hadrien Reynaud, Mengyun Qiao, Mischa Dombrowski, Thomas Day, Reza
Razavi, Alberto Gomez, Paul Leeson, Bernhard Kainz
- Abstract summary: We extend elucidated diffusion models for video modelling to generate plausible video sequences from single images.
Our image to sequence approach achieves an $R2$ score of 93%, 38 points higher than recently proposed sequence to sequence generation methods.
- Score: 5.102090025931326
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image synthesis is expected to provide value for the translation of machine
learning methods into clinical practice. Fundamental problems like model
robustness, domain transfer, causal modelling, and operator training become
approachable through synthetic data. Especially, heavily operator-dependant
modalities like Ultrasound imaging require robust frameworks for image and
video generation. So far, video generation has only been possible by providing
input data that is as rich as the output data, e.g., image sequence plus
conditioning in, video out. However, clinical documentation is usually scarce
and only single images are reported and stored, thus retrospective
patient-specific analysis or the generation of rich training data becomes
impossible with current approaches. In this paper, we extend elucidated
diffusion models for video modelling to generate plausible video sequences from
single images and arbitrary conditioning with clinical parameters. We explore
this idea within the context of echocardiograms by looking into the variation
of the Left Ventricle Ejection Fraction, the most essential clinical metric
gained from these examinations. We use the publicly available EchoNet-Dynamic
dataset for all our experiments. Our image to sequence approach achieves an
$R^2$ score of 93%, which is 38 points higher than recently proposed sequence
to sequence generation methods. Code and models will be available at:
https://github.com/HReynaud/EchoDiffusion.
Related papers
- Latent Drifting in Diffusion Models for Counterfactual Medical Image Synthesis [55.959002385347645]
Scaling by training on large datasets has been shown to enhance the quality and fidelity of image generation and manipulation with diffusion models.
Latent Drifting enables diffusion models to be conditioned for medical images fitted for the complex task of counterfactual image generation.
Our results demonstrate significant performance gains in various scenarios when combined with different fine-tuning schemes.
arXiv Detail & Related papers (2024-12-30T01:59:34Z) - SurgicaL-CD: Generating Surgical Images via Unpaired Image Translation with Latent Consistency Diffusion Models [1.6189876649941652]
We introduce emphSurgicaL-CD, a consistency-distilled diffusion method to generate realistic surgical images.
Our results demonstrate that our method outperforms GANs and diffusion-based approaches.
arXiv Detail & Related papers (2024-08-19T09:19:25Z) - EchoNet-Synthetic: Privacy-preserving Video Generation for Safe Medical Data Sharing [5.900946696794718]
We present a model designed to produce high-fidelity, long and accessible complete data samples with near-real-time efficiency.
We develop our generation method based on diffusion models and introduce a protocol for medical video dataset anonymization.
We present EchoNet-Synthetic, a fully synthetic, privacy-compliant echocardiogram dataset with paired ejection fraction labels.
arXiv Detail & Related papers (2024-06-02T17:18:06Z) - CathFlow: Self-Supervised Segmentation of Catheters in Interventional Ultrasound Using Optical Flow and Transformers [66.15847237150909]
We introduce a self-supervised deep learning architecture to segment catheters in longitudinal ultrasound images.
The network architecture builds upon AiAReSeg, a segmentation transformer built with the Attention in Attention mechanism.
We validated our model on a test dataset, consisting of unseen synthetic data and images collected from silicon aorta phantoms.
arXiv Detail & Related papers (2024-03-21T15:13:36Z) - DiffBoost: Enhancing Medical Image Segmentation via Text-Guided Diffusion Model [3.890243179348094]
Large-scale, big-variant, high-quality data are crucial for developing robust and successful deep-learning models for medical applications.
This paper proposes a novel approach by developing controllable diffusion models for medical image synthesis, called DiffBoost.
We leverage recent diffusion probabilistic models to generate realistic and diverse synthetic medical image data.
arXiv Detail & Related papers (2023-10-19T16:18:02Z) - D'ARTAGNAN: Counterfactual Video Generation [3.4079278794252232]
Causally-enabled machine learning frameworks could help clinicians to identify the best course of treatments by answering counterfactual questions.
We combine deep neural networks, twin causal networks and generative adversarial methods for the first time to build D'ARTAGNAN.
We generate new ultrasound videos, retaining the video style and anatomy of the original patient, with variations of the Ejection Fraction conditioned on a given input.
arXiv Detail & Related papers (2022-06-03T15:53:32Z) - Ultrasound Signal Processing: From Models to Deep Learning [64.56774869055826]
Medical ultrasound imaging relies heavily on high-quality signal processing to provide reliable and interpretable image reconstructions.
Deep learning based methods, which are optimized in a data-driven fashion, have gained popularity.
A relatively new paradigm combines the power of the two: leveraging data-driven deep learning, as well as exploiting domain knowledge.
arXiv Detail & Related papers (2022-04-09T13:04:36Z) - ViViT: A Video Vision Transformer [75.74690759089529]
We present pure-transformer based models for video classification.
Our model extracts-temporal tokens from the input video, which are then encoded by a series of transformer layers.
We show how we can effectively regularise the model during training and leverage pretrained image models to be able to train on comparatively small datasets.
arXiv Detail & Related papers (2021-03-29T15:27:17Z) - Image Translation for Medical Image Generation -- Ischemic Stroke
Lesions [0.0]
Synthetic databases with annotated pathologies could provide the required amounts of training data.
We train different image-to-image translation models to synthesize magnetic resonance images of brain volumes with and without stroke lesions.
We show that for a small database of only 10 or 50 clinical cases, synthetic data augmentation yields significant improvement.
arXiv Detail & Related papers (2020-10-05T09:12:28Z) - Towards Unsupervised Learning for Instrument Segmentation in Robotic
Surgery with Cycle-Consistent Adversarial Networks [54.00217496410142]
We propose an unpaired image-to-image translation where the goal is to learn the mapping between an input endoscopic image and a corresponding annotation.
Our approach allows to train image segmentation models without the need to acquire expensive annotations.
We test our proposed method on Endovis 2017 challenge dataset and show that it is competitive with supervised segmentation methods.
arXiv Detail & Related papers (2020-07-09T01:39:39Z) - Modeling Shared Responses in Neuroimaging Studies through MultiView ICA [94.31804763196116]
Group studies involving large cohorts of subjects are important to draw general conclusions about brain functional organization.
We propose a novel MultiView Independent Component Analysis model for group studies, where data from each subject are modeled as a linear combination of shared independent sources plus noise.
We demonstrate the usefulness of our approach first on fMRI data, where our model demonstrates improved sensitivity in identifying common sources among subjects.
arXiv Detail & Related papers (2020-06-11T17:29:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.