A Deep Learning Generative Model Approach for Image Synthesis of Plant
Leaves
- URL: http://arxiv.org/abs/2111.03388v1
- Date: Fri, 5 Nov 2021 10:53:35 GMT
- Title: A Deep Learning Generative Model Approach for Image Synthesis of Plant
Leaves
- Authors: Alessandrop Benfenati and Davide Bolzi and Paola Causin and Roberto
Oberti
- Abstract summary: We generate via advanced Deep Learning (DL) techniques artificial leaf images in an automatized way.
We aim to dispose of a source of training samples for AI applications for modern crop management.
- Score: 62.997667081978825
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Objectives. We generate via advanced Deep Learning (DL) techniques artificial
leaf images in an automatized way. We aim to dispose of a source of training
samples for AI applications for modern crop management. Such applications
require large amounts of data and, while leaf images are not truly scarce,
image collection and annotation remains a very time--consuming process. Data
scarcity can be addressed by augmentation techniques consisting in simple
transformations of samples belonging to a small dataset, but the richness of
the augmented data is limited: this motivates the search for alternative
approaches. Methods. Pursuing an approach based on DL generative models, we
propose a Leaf-to-Leaf Translation (L2L) procedure structured in two steps:
first, a residual variational autoencoder architecture generates synthetic leaf
skeletons (leaf profile and veins) starting from companions binarized skeletons
of real images. In a second step, we perform translation via a Pix2pix
framework, which uses conditional generator adversarial networks to reproduce
the colorization of leaf blades, preserving the shape and the venation pattern.
Results. The L2L procedure generates synthetic images of leaves with a
realistic appearance. We address the performance measurement both in a
qualitative and a quantitative way; for this latter evaluation, we employ a DL
anomaly detection strategy which quantifies the degree of anomaly of synthetic
leaves with respect to real samples. Conclusions. Generative DL approaches have
the potential to be a new paradigm to provide low-cost meaningful synthetic
samples for computer-aided applications. The present L2L approach represents a
step towards this goal, being able to generate synthetic samples with a
relevant qualitative and quantitative resemblance to real leaves.
Related papers
- Is Synthetic Image Useful for Transfer Learning? An Investigation into Data Generation, Volume, and Utilization [62.157627519792946]
We introduce a novel framework called bridged transfer, which initially employs synthetic images for fine-tuning a pre-trained model to improve its transferability.
We propose dataset style inversion strategy to improve the stylistic alignment between synthetic and real images.
Our proposed methods are evaluated across 10 different datasets and 5 distinct models, demonstrating consistent improvements.
arXiv Detail & Related papers (2024-03-28T22:25:05Z) - Bi-Modality Medical Image Synthesis Using Semi-Supervised Sequential
Generative Adversarial Networks [35.358653509217994]
We propose a bi-modality medical image synthesis approach based on sequential generative adversarial network (GAN) and semi-supervised learning.
Our approach consists of two generative modules that synthesize images of the two modalities in a sequential order.
Visual and quantitative results demonstrate the superiority of our method to the state-of-the-art methods.
arXiv Detail & Related papers (2023-08-27T10:39:33Z) - Explore the Power of Synthetic Data on Few-shot Object Detection [27.26215175101865]
Few-shot object detection (FSOD) aims to expand an object detector for novel categories given only a few instances for training.
Recent text-to-image generation models have shown promising results in generating high-quality images.
This work extensively studies how synthetic images generated from state-of-the-art text-to-image generators benefit FSOD tasks.
arXiv Detail & Related papers (2023-03-23T12:34:52Z) - Is synthetic data from generative models ready for image recognition? [69.42645602062024]
We study whether and how synthetic images generated from state-of-the-art text-to-image generation models can be used for image recognition tasks.
We showcase the powerfulness and shortcomings of synthetic data from existing generative models, and propose strategies for better applying synthetic data for recognition tasks.
arXiv Detail & Related papers (2022-10-14T06:54:24Z) - FewGAN: Generating from the Joint Distribution of a Few Images [95.6635227371479]
We introduce FewGAN, a generative model for generating novel, high-quality and diverse images.
FewGAN is a hierarchical patch-GAN that applies quantization at the first coarse scale, followed by a pyramid of residual fully convolutional GANs at finer scales.
In an extensive set of experiments, it is shown that FewGAN outperforms baselines both quantitatively and qualitatively.
arXiv Detail & Related papers (2022-07-18T07:11:28Z) - GLIDE: Towards Photorealistic Image Generation and Editing with
Text-Guided Diffusion Models [16.786221846896108]
We explore diffusion models for the problem of text-conditional image synthesis and compare two different guidance strategies.
We find that the latter is preferred by human evaluators for both photorealism and caption similarity, and often produces photorealistic samples.
Our models can be fine-tuned to perform image inpainting, enabling powerful text-driven image editing.
arXiv Detail & Related papers (2021-12-20T18:42:55Z) - METGAN: Generative Tumour Inpainting and Modality Synthesis in Light
Sheet Microscopy [4.872960046536882]
We introduce a novel generative method which leverages real anatomical information to generate realistic image-label pairs of tumours.
We construct a dual-pathway generator, for the anatomical image and label, trained in a cycle-consistent setup, constrained by an independent, pretrained segmentor.
The generated images yield significant quantitative improvement compared to existing methods.
arXiv Detail & Related papers (2021-04-22T11:18:17Z) - IMAGINE: Image Synthesis by Image-Guided Model Inversion [79.4691654458141]
We introduce an inversion based method, denoted as IMAge-Guided model INvErsion (IMAGINE), to generate high-quality and diverse images.
We leverage the knowledge of image semantics from a pre-trained classifier to achieve plausible generations.
IMAGINE enables the synthesis procedure to simultaneously 1) enforce semantic specificity constraints during the synthesis, 2) produce realistic images without generator training, and 3) give users intuitive control over the generation process.
arXiv Detail & Related papers (2021-04-13T02:00:24Z) - From Real to Synthetic and Back: Synthesizing Training Data for
Multi-Person Scene Understanding [0.7519872646378835]
We present a method for synthesizing naturally looking images of multiple people interacting in a specific scenario.
These images benefit from the advantages of synthetic data: being fully controllable and fully annotated with any type of standard or custom-defined ground truth.
To reduce the synthetic-to-real domain gap, we introduce a pipeline consisting of the following steps.
arXiv Detail & Related papers (2020-06-03T09:02:06Z) - High-Fidelity Synthesis with Disentangled Representation [60.19657080953252]
We propose an Information-Distillation Generative Adrial Network (ID-GAN) for disentanglement learning and high-fidelity synthesis.
Our method learns disentangled representation using VAE-based models, and distills the learned representation with an additional nuisance variable to the separate GAN-based generator for high-fidelity synthesis.
Despite the simplicity, we show that the proposed method is highly effective, achieving comparable image generation quality to the state-of-the-art methods using the disentangled representation.
arXiv Detail & Related papers (2020-01-13T14:39:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.