Controllable Data Generation Via Iterative Data-Property Mutual Mappings
- URL: http://arxiv.org/abs/2310.07683v1
- Date: Wed, 11 Oct 2023 17:34:56 GMT
- Title: Controllable Data Generation Via Iterative Data-Property Mutual Mappings
- Authors: Bo Pan, Muran Qin, Shiyu Wang, Yifei Zhang, Liang Zhao
- Abstract summary: We propose a framework to enhance VAE-based data generators with property controllability and ensure disentanglement.
The proposed framework is implemented on four VAE-based controllable generators to evaluate its performance on property error, disentanglement, generation quality, and training time.
- Score: 13.282793266390316
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep generative models have been widely used for their ability to generate
realistic data samples in various areas, such as images, molecules, text, and
speech. One major goal of data generation is controllability, namely to
generate new data with desired properties. Despite growing interest in the area
of controllable generation, significant challenges still remain, including 1)
disentangling desired properties with unrelated latent variables, 2)
out-of-distribution property control, and 3) objective optimization for
out-of-distribution property control. To address these challenges, in this
paper, we propose a general framework to enhance VAE-based data generators with
property controllability and ensure disentanglement. Our proposed objective can
be optimized on both data seen and unseen in the training set. We propose a
training procedure to train the objective in a semi-supervised manner by
iteratively conducting mutual mappings between the data and properties. The
proposed framework is implemented on four VAE-based controllable generators to
evaluate its performance on property error, disentanglement, generation
quality, and training time. The results indicate that our proposed framework
enables more precise control over the properties of generated samples in a
short training time, ensuring the disentanglement and keeping the validity of
the generated samples.
Related papers
- Generating the Traces You Need: A Conditional Generative Model for Process Mining Data [10.914597458295248]
We introduce a conditional model for process data generation based on a conditional variational autoencoder (CVAE)
CVAE for process mining faces specific challenges due to the multiperspective nature of the data and the need to adhere to control-flow rules.
arXiv Detail & Related papers (2024-11-04T14:44:20Z) - PerlDiff: Controllable Street View Synthesis Using Perspective-Layout Diffusion Models [55.080748327139176]
PerlDiff is a method for effective street view image generation that fully leverages perspective 3D geometric information.
Our results justify that our PerlDiff markedly enhances the precision of generation on the NuScenes and KITTI datasets.
arXiv Detail & Related papers (2024-07-08T16:46:47Z) - Implicitly Guided Design with PropEn: Match your Data to Follow the Gradient [52.2669490431145]
PropEn is inspired by'matching', which enables implicit guidance without training a discriminator.
We show that training with a matched dataset approximates the gradient of the property of interest while remaining within the data distribution.
arXiv Detail & Related papers (2024-05-28T11:30:19Z) - An improved tabular data generator with VAE-GMM integration [9.4491536689161]
We propose a novel Variational Autoencoder (VAE)-based model that addresses limitations of current approaches.
Inspired by the TVAE model, our approach incorporates a Bayesian Gaussian Mixture model (BGM) within the VAE architecture.
We thoroughly validate our model on three real-world datasets with mixed data types, including two medically relevant ones.
arXiv Detail & Related papers (2024-04-12T12:31:06Z) - SubjectDrive: Scaling Generative Data in Autonomous Driving via Subject Control [59.20038082523832]
We present SubjectDrive, the first model proven to scale generative data production in a way that could continuously improve autonomous driving applications.
We develop a novel model equipped with a subject control mechanism, which allows the generative model to leverage diverse external data sources for producing varied and useful data.
arXiv Detail & Related papers (2024-03-28T14:07:13Z) - Targeted Analysis of High-Risk States Using an Oriented Variational
Autoencoder [3.494548275937873]
Variational autoencoder (VAE) neural networks can be trained to generate power system states.
The coordinates of the latent space codes of VAEs have been shown to correlate with conceptual features of the data.
In this paper, an oriented variation autoencoder (OVAE) is proposed to constrain the link between latent space code and generated data.
arXiv Detail & Related papers (2023-03-20T19:34:21Z) - Multi-objective Deep Data Generation with Correlated Property Control [23.99970130388449]
We propose a novel deep generative framework that recovers semantics and the correlation of properties through disentangled latent vectors.
Our generative model preserves properties of interest while handling correlation and conflicts of properties under a multi-objective optimization framework.
arXiv Detail & Related papers (2022-10-01T00:35:45Z) - Attentive Prototypes for Source-free Unsupervised Domain Adaptive 3D
Object Detection [85.11649974840758]
3D object detection networks tend to be biased towards the data they are trained on.
We propose a single-frame approach for source-free, unsupervised domain adaptation of lidar-based 3D object detectors.
arXiv Detail & Related papers (2021-11-30T18:42:42Z) - TimeVAE: A Variational Auto-Encoder for Multivariate Time Series
Generation [6.824692201913679]
We propose a novel architecture for synthetically generating time-series data with the use of Variversaational Auto-Encoders (VAEs)
The proposed architecture has several distinct properties: interpretability, ability to encode domain knowledge, and reduced training times.
arXiv Detail & Related papers (2021-11-15T21:42:14Z) - Negative Data Augmentation [127.28042046152954]
We show that negative data augmentation samples provide information on the support of the data distribution.
We introduce a new GAN training objective where we use NDA as an additional source of synthetic data for the discriminator.
Empirically, models trained with our method achieve improved conditional/unconditional image generation along with improved anomaly detection capabilities.
arXiv Detail & Related papers (2021-02-09T20:28:35Z) - BREEDS: Benchmarks for Subpopulation Shift [98.90314444545204]
We develop a methodology for assessing the robustness of models to subpopulation shift.
We leverage the class structure underlying existing datasets to control the data subpopulations that comprise the training and test distributions.
Applying this methodology to the ImageNet dataset, we create a suite of subpopulation shift benchmarks of varying granularity.
arXiv Detail & Related papers (2020-08-11T17:04:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.