Applying generative neural networks for fast simulations of the ALICE (CERN) experiment
- URL: http://arxiv.org/abs/2407.16704v1
- Date: Wed, 10 Jul 2024 17:08:59 GMT
- Title: Applying generative neural networks for fast simulations of the ALICE (CERN) experiment
- Authors: Maksymilian Wojnar,
- Abstract summary: This thesis investigates the application of state-of-the-art advances in generative neural networks for fast simulation of the Zero Degree Calorimeter (ZDC) neutron detector at CERN.
Traditional simulation methods using the GEANT Monte Carlo toolkit, while accurate, are computationally demanding.
The thesis provides a comprehensive literature review on the application of neural networks in computer vision, fast simulations using machine learning, and generative neural networks in high-energy physics.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This thesis investigates the application of state-of-the-art advances in generative neural networks for fast simulation of the Zero Degree Calorimeter (ZDC) neutron detector in the ALICE experiment at CERN. Traditional simulation methods using the GEANT Monte Carlo toolkit, while accurate, are computationally demanding. With increasing computational needs at CERN, efficient simulation techniques are essential. The thesis provides a comprehensive literature review on the application of neural networks in computer vision, fast simulations using machine learning, and generative neural networks in high-energy physics. The theory of the analyzed models is also discussed, along with technical aspects and the challenges associated with a practical implementation. The experiments evaluate various neural network architectures, including convolutional neural networks, vision transformers, and MLP-Mixers, as well as generative frameworks such as autoencoders, generative adversarial networks, vector quantization models, and diffusion models. Key contributions include the implementation and evaluation of these models, a significant improvement in the Wasserstein metric compared to existing methods with a low generation time of 5 milliseconds per sample, and the formulation of a list of recommendations for developing models for fast ZDC simulation. Open-source code and detailed hyperparameter settings are provided for reproducibility. Additionally, the thesis outlines future research directions to further enhance simulation fidelity and efficiency.
Related papers
- A Realistic Simulation Framework for Analog/Digital Neuromorphic Architectures [73.65190161312555]
ARCANA is a spiking neural network simulator designed to account for the properties of mixed-signal neuromorphic circuits.
We show how the results obtained provide a reliable estimate of the behavior of the spiking neural network trained in software.
arXiv Detail & Related papers (2024-09-23T11:16:46Z) - Liquid Fourier Latent Dynamics Networks for fast GPU-based numerical simulations in computational cardiology [0.0]
We propose an extension of Latent Dynamics Networks (LDNets) to create parameterized space-time surrogate models for multiscale and multiphysics sets of highly nonlinear differential equations on complex geometries.
LFLDNets employ a neurologically-inspired, sparse liquid neural network for temporal dynamics, relaxing the requirement of a numerical solver for time advancement and leading to superior performance in terms of parameters, accuracy, efficiency and learned trajectories.
arXiv Detail & Related papers (2024-08-19T09:14:25Z) - Mechanistic Neural Networks for Scientific Machine Learning [58.99592521721158]
We present Mechanistic Neural Networks, a neural network design for machine learning applications in the sciences.
It incorporates a new Mechanistic Block in standard architectures to explicitly learn governing differential equations as representations.
Central to our approach is a novel Relaxed Linear Programming solver (NeuRLP) inspired by a technique that reduces solving linear ODEs to solving linear programs.
arXiv Detail & Related papers (2024-02-20T15:23:24Z) - Machine Learning methods for simulating particle response in the Zero
Degree Calorimeter at the ALICE experiment, CERN [8.980453507536017]
Currently, over half of the computing power at CERN GRID is used to run High Energy Physics simulations.
The recent updates at the Large Hadron Collider (LHC) create the need for developing more efficient simulation methods.
We propose an alternative approach to the problem that leverages machine learning.
arXiv Detail & Related papers (2023-06-23T16:45:46Z) - Automatic Evolution of Machine-Learning based Quantum Dynamics with
Uncertainty Analysis [4.629634111796585]
The long short-term memory recurrent neural network (LSTM-RNN) models are used to simulate the long-time quantum dynamics.
This work builds an effective machine learning approach to simulate the dynamics evolution of open quantum systems.
arXiv Detail & Related papers (2022-05-07T08:53:55Z) - An advanced spatio-temporal convolutional recurrent neural network for
storm surge predictions [73.4962254843935]
We study the capability of artificial neural network models to emulate storm surge based on the storm track/size/intensity history.
This study presents a neural network model that can predict storm surge, informed by a database of synthetic storm simulations.
arXiv Detail & Related papers (2022-04-18T23:42:18Z) - Optimising hadronic collider simulations using amplitude neural networks [0.0]
We train neural network models on one-loop amplitudes from the NJet C++ library and interface them with the Sherpa Monte Carlo event generator.
We find excellent agreement in the distributions and a reduced total simulation time by a factor of thirty.
arXiv Detail & Related papers (2022-02-09T15:08:30Z) - Mapping and Validating a Point Neuron Model on Intel's Neuromorphic
Hardware Loihi [77.34726150561087]
We investigate the potential of Intel's fifth generation neuromorphic chip - Loihi'
Loihi is based on the novel idea of Spiking Neural Networks (SNNs) emulating the neurons in the brain.
We find that Loihi replicates classical simulations very efficiently and scales notably well in terms of both time and energy performance as the networks get larger.
arXiv Detail & Related papers (2021-09-22T16:52:51Z) - Deep Bayesian Active Learning for Accelerating Stochastic Simulation [74.58219903138301]
Interactive Neural Process (INP) is a deep active learning framework for simulations and with active learning approaches.
For active learning, we propose a novel acquisition function, Latent Information Gain (LIG), calculated in the latent space of NP based models.
The results demonstrate STNP outperforms the baselines in the learning setting and LIG achieves the state-of-the-art for active learning.
arXiv Detail & Related papers (2021-06-05T01:31:51Z) - Invertible Surrogate Models: Joint surrogate modelling and
reconstruction of Laser-Wakefield Acceleration by invertible neural networks [55.41644538483948]
Invertible neural networks are a recent technique in machine learning.
We will be introducing invertible surrogate models that approximate complex forward simulation of the physics involved in laser plasma accelerators: iLWFA.
arXiv Detail & Related papers (2021-06-01T12:26:10Z) - Adaptive Neural Network-Based Approximation to Accelerate Eulerian Fluid
Simulation [9.576796509480445]
We introduce Smartfluidnet, a framework that automates model generation and application.
Smartfluidnet generates multiple neural networks before the simulation to meet the execution time and simulation quality requirement.
We show that Smartfluidnet achieves 1.46x and 590x speedup compared with a state-of-the-art neural network model and the original fluid simulation respectively.
arXiv Detail & Related papers (2020-08-26T21:44:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.