DODA: Diffusion for Object-detection Domain Adaptation in Agriculture
- URL: http://arxiv.org/abs/2403.18334v1
- Date: Wed, 27 Mar 2024 08:16:33 GMT
- Title: DODA: Diffusion for Object-detection Domain Adaptation in Agriculture
- Authors: Shuai Xiang, Pieter M. Blok, James Burridge, Haozhou Wang, Wei Guo,
- Abstract summary: We propose DODA, a data synthesizer that can generate high-quality object detection data for new domains in agriculture.
Specifically, we improve the controllability of layout-to-image through encoding layout as an image, thereby improving the quality of labels.
- Score: 4.549305421261851
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The diverse and high-quality content generated by recent generative models demonstrates the great potential of using synthetic data to train downstream models. However, in vision, especially in objection detection, related areas are not fully explored, the synthetic images are merely used to balance the long tails of existing datasets, and the accuracy of the generated labels is low, the full potential of generative models has not been exploited. In this paper, we propose DODA, a data synthesizer that can generate high-quality object detection data for new domains in agriculture. Specifically, we improve the controllability of layout-to-image through encoding layout as an image, thereby improving the quality of labels, and use a visual encoder to provide visual clues for the diffusion model to decouple visual features from the diffusion model, and empowering the model the ability to generate data in new domains. On the Global Wheat Head Detection (GWHD) Dataset, which is the largest dataset in agriculture and contains diverse domains, using the data synthesized by DODA improves the performance of the object detector by 12.74-17.76 AP$_{50}$ in the domain that was significantly shifted from the training data.
Related papers
- AeroGen: Enhancing Remote Sensing Object Detection with Diffusion-Driven Data Generation [38.89367726721828]
Remote sensing image object detection (RSIOD) aims to identify and locate specific objects within satellite or aerial imagery.
There is a scarcity of labeled data in current RSIOD datasets, which significantly limits the performance of current detection algorithms.
This paper proposes a layout-controllable diffusion generative model (i.e. AeroGen) tailored for RSIOD.
arXiv Detail & Related papers (2024-11-23T09:04:33Z) - Controlling Human Shape and Pose in Text-to-Image Diffusion Models via Domain Adaptation [1.3654846342364308]
We present a methodology for conditional control of human shape and pose in pretrained text-to-image diffusion models.
Fine-tuning these diffusion models to adhere to new conditions requires large datasets and high-quality annotations.
We propose a domain-adaptation technique that maintains image quality by isolating synthetically trained conditional information.
arXiv Detail & Related papers (2024-11-07T14:02:41Z) - From Web Data to Real Fields: Low-Cost Unsupervised Domain Adaptation for Agricultural Robots [3.7619101673213664]
This paper aims to adapt to specific fields at low cost using Unsupervised Domain Adaptation (UDA)
We explore a novel domain shift from a diverse, large pool of internet-sourced data to a small set of data collected by a robot at specific locations.
We introduce a novel module -- the Multi-level Attention-based Adrial Discriminator (MAAD) -- which can be integrated at the feature extractor level of any detection model.
arXiv Detail & Related papers (2024-10-31T13:11:09Z) - A Simple Background Augmentation Method for Object Detection with Diffusion Model [53.32935683257045]
In computer vision, it is well-known that a lack of data diversity will impair model performance.
We propose a simple yet effective data augmentation approach by leveraging advancements in generative models.
Background augmentation, in particular, significantly improves the models' robustness and generalization capabilities.
arXiv Detail & Related papers (2024-08-01T07:40:00Z) - ADLDA: A Method to Reduce the Harm of Data Distribution Shift in Data Augmentation [11.887799310374174]
This study introduces a novel data augmentation technique, ADLDA, aimed at mitigating the negative impact of data distribution shifts.
Experimental results demonstrate that ADLDA significantly enhances model performance across multiple datasets.
arXiv Detail & Related papers (2024-05-11T03:20:35Z) - DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception [78.26734070960886]
Current perceptive models heavily depend on resource-intensive datasets.
We introduce perception-aware loss (P.A. loss) through segmentation, improving both quality and controllability.
Our method customizes data augmentation by extracting and utilizing perception-aware attribute (P.A. Attr) during generation.
arXiv Detail & Related papers (2024-03-20T04:58:03Z) - DreamDA: Generative Data Augmentation with Diffusion Models [68.22440150419003]
This paper proposes a new classification-oriented framework DreamDA.
DreamDA generates diverse samples that adhere to the original data distribution by considering training images in the original data as seeds.
In addition, since the labels of the generated data may not align with the labels of their corresponding seed images, we introduce a self-training paradigm for generating pseudo labels.
arXiv Detail & Related papers (2024-03-19T15:04:35Z) - GenFace: A Large-Scale Fine-Grained Face Forgery Benchmark and Cross Appearance-Edge Learning [50.7702397913573]
The rapid advancement of photorealistic generators has reached a critical juncture where the discrepancy between authentic and manipulated images is increasingly indistinguishable.
Although there have been a number of publicly available face forgery datasets, the forgery faces are mostly generated using GAN-based synthesis technology.
We propose a large-scale, diverse, and fine-grained high-fidelity dataset, namely GenFace, to facilitate the advancement of deepfake detection.
arXiv Detail & Related papers (2024-02-03T03:13:50Z) - UAV-Sim: NeRF-based Synthetic Data Generation for UAV-based Perception [62.71374902455154]
We leverage recent advancements in neural rendering to improve static and dynamic novelview UAV-based image rendering.
We demonstrate a considerable performance boost when a state-of-the-art detection model is optimized primarily on hybrid sets of real and synthetic data.
arXiv Detail & Related papers (2023-10-25T00:20:37Z) - Generative Modeling Helps Weak Supervision (and Vice Versa) [87.62271390571837]
We propose a model fusing weak supervision and generative adversarial networks.
It captures discrete variables in the data alongside the weak supervision derived label estimate.
It is the first approach to enable data augmentation through weakly supervised synthetic images and pseudolabels.
arXiv Detail & Related papers (2022-03-22T20:24:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.