Hierarchically branched diffusion models leverage dataset structure for
class-conditional generation
- URL: http://arxiv.org/abs/2212.10777v4
- Date: Thu, 1 Feb 2024 23:34:04 GMT
- Title: Hierarchically branched diffusion models leverage dataset structure for
class-conditional generation
- Authors: Alex M. Tseng, Max Shen, Tommaso Biancalani, Gabriele Scalia
- Abstract summary: Branched diffusion models rely on the same diffusion process as traditional models, but learn reverse diffusion separately for each branch of a hierarchy.
We extensively evaluate branched diffusion models on several benchmark and large real-world scientific datasets spanning many data modalities.
- Score: 0.6800113478497425
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Class-labeled datasets, particularly those common in scientific domains, are
rife with internal structure, yet current class-conditional diffusion models
ignore these relationships and implicitly diffuse on all classes in a flat
fashion. To leverage this structure, we propose hierarchically branched
diffusion models as a novel framework for class-conditional generation.
Branched diffusion models rely on the same diffusion process as traditional
models, but learn reverse diffusion separately for each branch of a hierarchy.
We highlight several advantages of branched diffusion models over the current
state-of-the-art methods for class-conditional diffusion, including extension
to novel classes in a continual-learning setting, a more sophisticated form of
analogy-based conditional generation (i.e. transmutation), and a novel
interpretability into the generation process. We extensively evaluate branched
diffusion models on several benchmark and large real-world scientific datasets
spanning many data modalities.
Related papers
- How compositional generalization and creativity improve as diffusion models are trained [82.08869888944324]
How many samples do generative models need to learn the composition rules, so as to produce a number of novel data?
We consider diffusion models trained on simple context-free grammars - tree-like graphical models used to represent the structure of data such as language and images.
We demonstrate that diffusion models learn compositional rules with the sample complexity required for clustering features with statistically similar context, a process similar to the word2vec.
arXiv Detail & Related papers (2025-02-17T18:06:33Z) - Continuous Diffusion Model for Language Modeling [57.396578974401734]
Existing continuous diffusion models for discrete data have limited performance compared to discrete approaches.
We propose a continuous diffusion model for language modeling that incorporates the geometry of the underlying categorical distribution.
arXiv Detail & Related papers (2025-02-17T08:54:29Z) - On the Feature Learning in Diffusion Models [26.53807235141923]
We propose a feature learning framework aimed at analyzing and comparing the training dynamics of diffusion models with those of traditional classification models.
Our theoretical analysis demonstrates that, under identical settings, diffusion models are encouraged to learn more balanced and comprehensive representations of the data.
In contrast, neural networks with a similar architecture trained for classification tend to prioritize learning specific patterns in the data.
arXiv Detail & Related papers (2024-12-02T00:41:25Z) - Unveil Conditional Diffusion Models with Classifier-free Guidance: A Sharp Statistical Theory [87.00653989457834]
Conditional diffusion models serve as the foundation of modern image synthesis and find extensive application in fields like computational biology and reinforcement learning.
Despite the empirical success, theory of conditional diffusion models is largely missing.
This paper bridges the gap by presenting a sharp statistical theory of distribution estimation using conditional diffusion models.
arXiv Detail & Related papers (2024-03-18T17:08:24Z) - A Phase Transition in Diffusion Models Reveals the Hierarchical Nature of Data [51.03144354630136]
Recent advancements show that diffusion models can generate high-quality images.
We study this phenomenon in a hierarchical generative model of data.
We find that the backward diffusion process acting after a time $t$ is governed by a phase transition.
arXiv Detail & Related papers (2024-02-26T19:52:33Z) - Training Class-Imbalanced Diffusion Model Via Overlap Optimization [55.96820607533968]
Diffusion models trained on real-world datasets often yield inferior fidelity for tail classes.
Deep generative models, including diffusion models, are biased towards classes with abundant training images.
We propose a method based on contrastive learning to minimize the overlap between distributions of synthetic images for different classes.
arXiv Detail & Related papers (2024-02-16T16:47:21Z) - Navigating the Structured What-If Spaces: Counterfactual Generation via
Structured Diffusion [20.20945739504847]
We introduce Structured Counterfactual diffuser or SCD, the first plug-and-play framework leveraging diffusion for generating counterfactual explanations in structured data.
Our experiments show that our counterfactuals not only exhibit high plausibility compared to the existing state-of-the-art but also show significantly better proximity and diversity.
arXiv Detail & Related papers (2023-12-21T07:05:21Z) - The Emergence of Reproducibility and Generalizability in Diffusion Models [10.188731323681575]
Given the same starting noise input and a deterministic sampler, different diffusion models often yield remarkably similar outputs.
We show that diffusion models are learning distinct distributions affected by the training data size.
This valuable property generalizes to many variants of diffusion models, including those for conditional use, solving inverse problems, and model fine-tuning.
arXiv Detail & Related papers (2023-10-08T19:02:46Z) - Renormalizing Diffusion Models [0.7252027234425334]
We use diffusion models to learn inverse renormalization group flows of statistical and quantum field theories.
Our work provides an interpretation of multiscale diffusion models, and gives physically-inspired suggestions for diffusion models which should have novel properties.
arXiv Detail & Related papers (2023-08-23T18:02:31Z) - A Survey on Generative Diffusion Model [75.93774014861978]
Diffusion models are an emerging class of deep generative models.
They have certain limitations, including a time-consuming iterative generation process and confinement to high-dimensional Euclidean space.
This survey presents a plethora of advanced techniques aimed at enhancing diffusion models.
arXiv Detail & Related papers (2022-09-06T16:56:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.