LC-GAN: Image-to-image Translation Based on Generative Adversarial
Network for Endoscopic Images
- URL: http://arxiv.org/abs/2003.04949v2
- Date: Thu, 13 Aug 2020 21:24:33 GMT
- Title: LC-GAN: Image-to-image Translation Based on Generative Adversarial
Network for Endoscopic Images
- Authors: Shan Lin, Fangbo Qin, Yangming Li, Randall A. Bly, Kris S. Moe, Blake
Hannaford
- Abstract summary: We propose an image-to-image translation model live-cadaver GAN (LC-GAN) based on generative adversarial networks (GANs)
For live image segmentation, we first translate the live images to fake-cadaveric images with LC-GAN and then perform segmentation on the fake-cadaveric images with models trained on the real cadaveric dataset.
Our model achieves better image-to-image translation and leads to improved segmentation performance in the proposed cross-domain segmentation task.
- Score: 22.253074722129053
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Intelligent vision is appealing in computer-assisted and robotic surgeries.
Vision-based analysis with deep learning usually requires large labeled
datasets, but manual data labeling is expensive and time-consuming in medical
problems. We investigate a novel cross-domain strategy to reduce the need for
manual data labeling by proposing an image-to-image translation model
live-cadaver GAN (LC-GAN) based on generative adversarial networks (GANs). We
consider a situation when a labeled cadaveric surgery dataset is available
while the task is instrument segmentation on an unlabeled live surgery dataset.
We train LC-GAN to learn the mappings between the cadaveric and live images.
For live image segmentation, we first translate the live images to
fake-cadaveric images with LC-GAN and then perform segmentation on the
fake-cadaveric images with models trained on the real cadaveric dataset. The
proposed method fully makes use of the labeled cadaveric dataset for live image
segmentation without the need to label the live dataset. LC-GAN has two
generators with different architectures that leverage the deep feature
representation learned from the cadaveric image based segmentation task.
Moreover, we propose the structural similarity loss and segmentation
consistency loss to improve the semantic consistency during translation. Our
model achieves better image-to-image translation and leads to improved
segmentation performance in the proposed cross-domain segmentation task.
Related papers
- Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - Dataset Diffusion: Diffusion-based Synthetic Dataset Generation for
Pixel-Level Semantic Segmentation [6.82236459614491]
We propose a novel method for generating pixel-level semantic segmentation labels using the text-to-image generative model Stable Diffusion.
By utilizing the text prompts, cross-attention, and self-attention of SD, we introduce three new techniques: class-prompt appending, class-prompt cross-attention, and self-attention exponentiation.
These techniques enable us to generate segmentation maps corresponding to synthetic images.
arXiv Detail & Related papers (2023-09-25T17:19:26Z) - Improving CT Image Segmentation Accuracy Using StyleGAN Driven Data
Augmentation [42.034896915716374]
This paper presents a StyleGAN-driven approach for segmenting publicly available large medical datasets.
Style transfer is used to augment the training dataset and generate new anatomically sound images.
The augmented dataset is then used to train a U-Net segmentation network which displays a significant improvement in the segmentation accuracy.
arXiv Detail & Related papers (2023-02-07T06:34:10Z) - Self-Supervised Correction Learning for Semi-Supervised Biomedical Image
Segmentation [84.58210297703714]
We propose a self-supervised correction learning paradigm for semi-supervised biomedical image segmentation.
We design a dual-task network, including a shared encoder and two independent decoders for segmentation and lesion region inpainting.
Experiments on three medical image segmentation datasets for different tasks demonstrate the outstanding performance of our method.
arXiv Detail & Related papers (2023-01-12T08:19:46Z) - Extracting Semantic Knowledge from GANs with Unsupervised Learning [65.32631025780631]
Generative Adversarial Networks (GANs) encode semantics in feature maps in a linearly separable form.
We propose a novel clustering algorithm, named KLiSH, which leverages the linear separability to cluster GAN's features.
KLiSH succeeds in extracting fine-grained semantics of GANs trained on datasets of various objects.
arXiv Detail & Related papers (2022-11-30T03:18:16Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Enhancing MR Image Segmentation with Realistic Adversarial Data
Augmentation [17.539828821476224]
We propose an adversarial data augmentation approach to improve the efficiency in utilizing training data.
We present a generic task-driven learning framework, which jointly optimize a data augmentation model and a segmentation network during training.
The proposed adversarial data augmentation does not rely on generative networks and can be used as a plug-in module in general segmentation networks.
arXiv Detail & Related papers (2021-08-07T11:32:37Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Pairwise Relation Learning for Semi-supervised Gland Segmentation [90.45303394358493]
We propose a pairwise relation-based semi-supervised (PRS2) model for gland segmentation on histology images.
This model consists of a segmentation network (S-Net) and a pairwise relation network (PR-Net)
We evaluate our model against five recent methods on the GlaS dataset and three recent methods on the CRAG dataset.
arXiv Detail & Related papers (2020-08-06T15:02:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.