End-To-End Latent Variational Diffusion Models for Inverse Problems in
High Energy Physics
- URL: http://arxiv.org/abs/2305.10399v1
- Date: Wed, 17 May 2023 17:43:10 GMT
- Title: End-To-End Latent Variational Diffusion Models for Inverse Problems in
High Energy Physics
- Authors: Alexander Shmakov, Kevin Greif, Michael Fenton, Aishik Ghosh, Pierre
Baldi, Daniel Whiteson
- Abstract summary: We introduce a novel unified architecture, termed latent variation models, which combines the latent learning of cutting-edge generative art approaches with an end-to-end variational framework.
Our unified approach achieves a distribution-free distance to the truth of over 20 times less than non-latent state-of-the-art baseline.
- Score: 61.44793171735013
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: High-energy collisions at the Large Hadron Collider (LHC) provide valuable
insights into open questions in particle physics. However, detector effects
must be corrected before measurements can be compared to certain theoretical
predictions or measurements from other detectors. Methods to solve this
\textit{inverse problem} of mapping detector observations to theoretical
quantities of the underlying collision are essential parts of many physics
analyses at the LHC. We investigate and compare various generative deep
learning methods to approximate this inverse mapping. We introduce a novel
unified architecture, termed latent variation diffusion models, which combines
the latent learning of cutting-edge generative art approaches with an
end-to-end variational framework. We demonstrate the effectiveness of this
approach for reconstructing global distributions of theoretical kinematic
quantities, as well as for ensuring the adherence of the learned posterior
distributions to known physics constraints. Our unified approach achieves a
distribution-free distance to the truth of over 20 times less than non-latent
state-of-the-art baseline and 3 times less than traditional latent diffusion
models.
Related papers
- Neural Message Passing Induced by Energy-Constrained Diffusion [79.9193447649011]
We propose an energy-constrained diffusion model as a principled interpretable framework for understanding the mechanism of MPNNs.
We show that the new model can yield promising performance for cases where the data structures are observed (as a graph), partially observed or completely unobserved.
arXiv Detail & Related papers (2024-09-13T17:54:41Z) - Variational Pseudo Marginal Methods for Jet Reconstruction in Particle Physics [2.223804777595989]
We introduce a Combinatorial Sequential Monte Carlo approach for inferring jet latent structures.
As a second contribution, we leverage the resulting estimator to develop a variational inference algorithm for parameter learning.
We illustrate our method's effectiveness through experiments using data generated with a collider physics generative model.
arXiv Detail & Related papers (2024-06-05T13:18:55Z) - Deciphering diffuse scattering with machine learning and the equivariant
foundation model: The case of molten FeO [5.643890195191516]
Bridging the gap between diffuse x-ray or neutron scattering measurements and predicted structures from atom-atom pair potentials in disordered materials has been a longstanding challenge in condensed matter physics.
The use of machine learned interatomic potentials has grown in the past few years, and has been particularly successful in the cases of ionic and oxide systems.
Here we leverage MACE-MP-0; a newly introduced equivariant foundation model and validate the results against high-quality experimental scattering data for the case of molten iron(II) oxide (FeO)
arXiv Detail & Related papers (2024-03-01T03:50:03Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - Energy Discrepancies: A Score-Independent Loss for Energy-Based Models [20.250792836049882]
We propose a novel loss function called Energy Discrepancy (ED) which does not rely on the computation of scores or expensive Markov chain Monte Carlo.
We show that ED approaches the explicit score matching and negative log-likelihood loss under different limits, effectively interpolating between both.
arXiv Detail & Related papers (2023-07-12T19:51:49Z) - Diffusion Models are Minimax Optimal Distribution Estimators [49.47503258639454]
We provide the first rigorous analysis on approximation and generalization abilities of diffusion modeling.
We show that when the true density function belongs to the Besov space and the empirical score matching loss is properly minimized, the generated data distribution achieves the nearly minimax optimal estimation rates.
arXiv Detail & Related papers (2023-03-03T11:31:55Z) - Let us Build Bridges: Understanding and Extending Diffusion Generative
Models [19.517597928769042]
Diffusion-based generative models have achieved promising results recently, but raise an array of open questions.
This work tries to re-exam the overall framework in order to gain better theoretical understandings.
We present 1) a first theoretical error analysis for learning diffusion generation models, and 2) a simple and unified approach to learning on data from different discrete and constrained domains.
arXiv Detail & Related papers (2022-08-31T08:58:10Z) - How Much is Enough? A Study on Diffusion Times in Score-based Generative
Models [76.76860707897413]
Current best practice advocates for a large T to ensure that the forward dynamics brings the diffusion sufficiently close to a known and simple noise distribution.
We show how an auxiliary model can be used to bridge the gap between the ideal and the simulated forward dynamics, followed by a standard reverse diffusion process.
arXiv Detail & Related papers (2022-06-10T15:09:46Z) - Quantifying Model Predictive Uncertainty with Perturbation Theory [21.591460685054546]
We propose a framework for predictive uncertainty quantification of a neural network.
We use perturbation theory from quantum physics to formulate a moment decomposition problem.
Our approach provides fast model predictive uncertainty estimates with much greater precision and calibration.
arXiv Detail & Related papers (2021-09-22T17:55:09Z) - Targeted free energy estimation via learned mappings [66.20146549150475]
Free energy perturbation (FEP) was proposed by Zwanzig more than six decades ago as a method to estimate free energy differences.
FEP suffers from a severe limitation: the requirement of sufficient overlap between distributions.
One strategy to mitigate this problem, called Targeted Free Energy Perturbation, uses a high-dimensional mapping in configuration space to increase overlap.
arXiv Detail & Related papers (2020-02-12T11:10:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.