Improving Generative Model-based Unfolding with Schr\"{o}dinger Bridges
- URL: http://arxiv.org/abs/2308.12351v2
- Date: Fri, 22 Sep 2023 17:28:21 GMT
- Title: Improving Generative Model-based Unfolding with Schr\"{o}dinger Bridges
- Authors: Sascha Diefenbacher, Guan-Horng Liu, Vinicius Mikuni, Benjamin
Nachman, and Weili Nie
- Abstract summary: Machine learning-based unfolding has enabled unbinned and high-dimensional differential cross section measurements.
We propose to use Schroedinger Bridges and diffusion models to create SBUnfold, an unfolding approach that combines the strengths of both discriminative and generative models.
We show that SBUnfold achieves excellent performance compared to state of the art methods on a synthetic Z+jets dataset.
- Score: 14.989614554242229
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning-based unfolding has enabled unbinned and high-dimensional
differential cross section measurements. Two main approaches have emerged in
this research area: one based on discriminative models and one based on
generative models. The main advantage of discriminative models is that they
learn a small correction to a starting simulation while generative models scale
better to regions of phase space with little data. We propose to use
Schroedinger Bridges and diffusion models to create SBUnfold, an unfolding
approach that combines the strengths of both discriminative and generative
models. The key feature of SBUnfold is that its generative model maps one set
of events into another without having to go through a known probability density
as is the case for normalizing flows and standard diffusion models. We show
that SBUnfold achieves excellent performance compared to state of the art
methods on a synthetic Z+jets dataset.
Related papers
- Generative Unfolding with Distribution Mapping [0.0837622912636323]
We show how to extend two morphing techniques, Schr"odinger Bridges and Direct Diffusion, in order to ensure that the models learn the correct conditional probabilities.
Results are presented with a standard benchmark dataset of single jet substructure as well as for a new dataset describing a 22-dimensional phase space of Z + 2-jets.
arXiv Detail & Related papers (2024-11-04T19:00:01Z) - Energy-Based Diffusion Language Models for Text Generation [126.23425882687195]
Energy-based Diffusion Language Model (EDLM) is an energy-based model operating at the full sequence level for each diffusion step.
Our framework offers a 1.3$times$ sampling speedup over existing diffusion models.
arXiv Detail & Related papers (2024-10-28T17:25:56Z) - Bridging Model-Based Optimization and Generative Modeling via Conservative Fine-Tuning of Diffusion Models [54.132297393662654]
We introduce a hybrid method that fine-tunes cutting-edge diffusion models by optimizing reward models through RL.
We demonstrate the capability of our approach to outperform the best designs in offline data, leveraging the extrapolation capabilities of reward models.
arXiv Detail & Related papers (2024-05-30T03:57:29Z) - Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution [67.9215891673174]
We propose score entropy as a novel loss that naturally extends score matching to discrete spaces.
We test our Score Entropy Discrete Diffusion models on standard language modeling tasks.
arXiv Detail & Related papers (2023-10-25T17:59:12Z) - Diff-Instruct: A Universal Approach for Transferring Knowledge From
Pre-trained Diffusion Models [77.83923746319498]
We propose a framework called Diff-Instruct to instruct the training of arbitrary generative models.
We show that Diff-Instruct results in state-of-the-art single-step diffusion-based models.
Experiments on refining GAN models show that the Diff-Instruct can consistently improve the pre-trained generators of GAN models.
arXiv Detail & Related papers (2023-05-29T04:22:57Z) - On Distillation of Guided Diffusion Models [94.95228078141626]
We propose an approach to distilling classifier-free guided diffusion models into models that are fast to sample from.
For standard diffusion models trained on the pixelspace, our approach is able to generate images visually comparable to that of the original model.
For diffusion models trained on the latent-space (e.g., Stable Diffusion), our approach is able to generate high-fidelity images using as few as 1 to 4 denoising steps.
arXiv Detail & Related papers (2022-10-06T18:03:56Z) - Model-free Subsampling Method Based on Uniform Designs [5.661822729320697]
We develop a low-GEFD data-driven subsampling method based on the existing uniform designs.
Our method keeps robust under diverse model specifications while other popular subsampling methods are under-performing.
arXiv Detail & Related papers (2022-09-08T07:47:56Z) - A Survey on Generative Diffusion Model [75.93774014861978]
Diffusion models are an emerging class of deep generative models.
They have certain limitations, including a time-consuming iterative generation process and confinement to high-dimensional Euclidean space.
This survey presents a plethora of advanced techniques aimed at enhancing diffusion models.
arXiv Detail & Related papers (2022-09-06T16:56:21Z) - Normalizing Flow based Hidden Markov Models for Classification of Speech
Phones with Explainability [25.543231171094384]
In pursuit of explainability, we develop generative models for sequential data.
We combine modern neural networks (normalizing flows) and traditional generative models (hidden Markov models - HMMs)
The proposed generative models can compute likelihood of a data and hence directly suitable for maximum-likelihood (ML) classification approach.
arXiv Detail & Related papers (2021-07-01T20:10:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.