Expanding Small-Scale Datasets with Guided Imagination
- URL: http://arxiv.org/abs/2211.13976v6
- Date: Tue, 10 Oct 2023 07:18:46 GMT
- Title: Expanding Small-Scale Datasets with Guided Imagination
- Authors: Yifan Zhang, Daquan Zhou, Bryan Hooi, Kai Wang, Jiashi Feng
- Abstract summary: dataset expansion is a new task aimed at expanding a ready-to-use small dataset by automatically creating new labeled samples.
GIF conducts data imagination by optimizing the latent features of the seed data in the semantically meaningful space of the prior model.
GIF-SD obtains 13.5% higher model accuracy on natural image datasets than unguided expansion with SD.
- Score: 92.5276783917845
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The power of DNNs relies heavily on the quantity and quality of training
data. However, collecting and annotating data on a large scale is often
expensive and time-consuming. To address this issue, we explore a new task,
termed dataset expansion, aimed at expanding a ready-to-use small dataset by
automatically creating new labeled samples. To this end, we present a Guided
Imagination Framework (GIF) that leverages cutting-edge generative models like
DALL-E2 and Stable Diffusion (SD) to "imagine" and create informative new data
from the input seed data. Specifically, GIF conducts data imagination by
optimizing the latent features of the seed data in the semantically meaningful
space of the prior model, resulting in the creation of photo-realistic images
with new content. To guide the imagination towards creating informative samples
for model training, we introduce two key criteria, i.e., class-maintained
information boosting and sample diversity promotion. These criteria are
verified to be essential for effective dataset expansion: GIF-SD obtains 13.5%
higher model accuracy on natural image datasets than unguided expansion with
SD. With these essential criteria, GIF successfully expands small datasets in
various scenarios, boosting model accuracy by 36.9% on average over six natural
image datasets and by 13.5% on average over three medical datasets. The source
code is available at https://github.com/Vanint/DatasetExpansion.
Related papers
- Erase, then Redraw: A Novel Data Augmentation Approach for Free Space Detection Using Diffusion Model [5.57325257338134]
Traditional data augmentation methods cannot alter high-level semantic attributes.
We propose a text-to-image diffusion model to parameterize image-to-image transformations.
We achieve this goal by erasing instances of real objects from the original dataset and generating new instances with similar semantics in the erased regions.
arXiv Detail & Related papers (2024-09-30T10:21:54Z) - 3D-VirtFusion: Synthetic 3D Data Augmentation through Generative Diffusion Models and Controllable Editing [52.68314936128752]
We propose a new paradigm to automatically generate 3D labeled training data by harnessing the power of pretrained large foundation models.
For each target semantic class, we first generate 2D images of a single object in various structure and appearance via diffusion models and chatGPT generated text prompts.
We transform these augmented images into 3D objects and construct virtual scenes by random composition.
arXiv Detail & Related papers (2024-08-25T09:31:22Z) - DataDream: Few-shot Guided Dataset Generation [90.09164461462365]
We propose a framework for synthesizing classification datasets that more faithfully represents the real data distribution.
DataDream fine-tunes LoRA weights for the image generation model on the few real images before generating the training data using the adapted model.
We then fine-tune LoRA weights for CLIP using the synthetic data to improve downstream image classification over previous approaches on a large variety of datasets.
arXiv Detail & Related papers (2024-07-15T17:10:31Z) - Distribution-Aware Data Expansion with Diffusion Models [55.979857976023695]
We propose DistDiff, a training-free data expansion framework based on the distribution-aware diffusion model.
DistDiff consistently enhances accuracy across a diverse range of datasets compared to models trained solely on original data.
arXiv Detail & Related papers (2024-03-11T14:07:53Z) - Retrieval-Augmented Data Augmentation for Low-Resource Domain Tasks [66.87070857705994]
In low-resource settings, the amount of seed data samples to use for data augmentation is very small.
We propose a novel method that augments training data by incorporating a wealth of examples from other datasets.
This approach can ensure that the generated data is not only relevant but also more diverse than what could be achieved using the limited seed data alone.
arXiv Detail & Related papers (2024-02-21T02:45:46Z) - DatasetDM: Synthesizing Data with Perception Annotations Using Diffusion
Models [61.906934570771256]
We present a generic dataset generation model that can produce diverse synthetic images and perception annotations.
Our method builds upon the pre-trained diffusion model and extends text-guided image synthesis to perception data generation.
We show that the rich latent code of the diffusion model can be effectively decoded as accurate perception annotations using a decoder module.
arXiv Detail & Related papers (2023-08-11T14:38:11Z) - T-ADAF: Adaptive Data Augmentation Framework for Image Classification
Network based on Tensor T-product Operator [0.0]
This paper proposes an Adaptive Data Augmentation Framework based on the tensor T-product Operator.
It triples one image data to be trained and gain the result from all these three images together with only less than 0.1% increase in the number of parameters.
Numerical experiments show that our data augmentation framework can improve the performance of original neural network model by 2%.
arXiv Detail & Related papers (2023-06-07T08:30:44Z) - Training on Thin Air: Improve Image Classification with Generated Data [28.96941414724037]
Diffusion Inversion is a simple yet effective method to generate diverse, high-quality training data for image classification.
Our approach captures the original data distribution and ensures data coverage by inverting images to the latent space of Stable Diffusion.
We identify three key components that allow our generated images to successfully supplant the original dataset.
arXiv Detail & Related papers (2023-05-24T16:33:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.