Explainable Equivariant Neural Networks for Particle Physics: PELICAN
- URL: http://arxiv.org/abs/2307.16506v4
- Date: Fri, 23 Feb 2024 20:27:25 GMT
- Title: Explainable Equivariant Neural Networks for Particle Physics: PELICAN
- Authors: Alexander Bogatskiy, Timothy Hoffman, David W. Miller, Jan T.
Offermann, Xiaoyang Liu
- Abstract summary: PELICAN is a novel permutation equivariant and Lorentz invariant aggregator network.
We present a study of the PELICAN algorithm architecture in the context of both tagging (classification) and reconstructing (regression) Lorentz-boosted top quarks.
We extend the application of PELICAN to the tasks of identifying quark-initiated vs.gluon-initiated jets, and a multi-class identification across five separate target categories of jets.
- Score: 51.02649432050852
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: PELICAN is a novel permutation equivariant and Lorentz invariant or covariant
aggregator network designed to overcome common limitations found in
architectures applied to particle physics problems. Compared to many approaches
that use non-specialized architectures that neglect underlying physics
principles and require very large numbers of parameters, PELICAN employs a
fundamentally symmetry group-based architecture that demonstrates benefits in
terms of reduced complexity, increased interpretability, and raw performance.
We present a comprehensive study of the PELICAN algorithm architecture in the
context of both tagging (classification) and reconstructing (regression)
Lorentz-boosted top quarks, including the difficult task of specifically
identifying and measuring the $W$-boson inside the dense environment of the
Lorentz-boosted top-quark hadronic final state. We also extend the application
of PELICAN to the tasks of identifying quark-initiated vs.~gluon-initiated
jets, and a multi-class identification across five separate target categories
of jets. When tested on the standard task of Lorentz-boosted top-quark tagging,
PELICAN outperforms existing competitors with much lower model complexity and
high sample efficiency. On the less common and more complex task of 4-momentum
regression, PELICAN also outperforms hand-crafted, non-machine learning
algorithms. We discuss the implications of symmetry-restricted architectures
for the wider field of machine learning for physics.
Related papers
- Lorentz-Equivariant Geometric Algebra Transformers for High-Energy Physics [4.4970885242855845]
Lorentz Geometric Algebra Transformer (L-GATr) is a new multi-purpose architecture for high-energy physics.
L-GATr is first demonstrated on regression and classification tasks from particle physics.
We then construct the first Lorentz-equivariant generative model: a continuous normalizing flow based on an L-GATr network.
arXiv Detail & Related papers (2024-05-23T17:15:41Z) - 19 Parameters Is All You Need: Tiny Neural Networks for Particle Physics [52.42485649300583]
We present the potential of one recent Lorentz- and permutation-symmetric architecture, PELICAN, for low-latency neural network tasks.
We show its instances with as few as 19 trainable parameters that outperform generic architectures with tens of thousands of parameters when compared on the binary classification task of top quark jet tagging.
arXiv Detail & Related papers (2023-10-24T18:51:22Z) - PELICAN: Permutation Equivariant and Lorentz Invariant or Covariant
Aggregator Network for Particle Physics [64.5726087590283]
We present a machine learning architecture that uses a set of inputs maximally reduced with respect to the full 6-dimensional Lorentz symmetry.
We show that the resulting network outperforms all existing competitors despite much lower model complexity.
arXiv Detail & Related papers (2022-11-01T13:36:50Z) - Tunable Complexity Benchmarks for Evaluating Physics-Informed Neural
Networks on Coupled Ordinary Differential Equations [64.78260098263489]
In this work, we assess the ability of physics-informed neural networks (PINNs) to solve increasingly-complex coupled ordinary differential equations (ODEs)
We show that PINNs eventually fail to produce correct solutions to these benchmarks as their complexity increases.
We identify several reasons why this may be the case, including insufficient network capacity, poor conditioning of the ODEs, and high local curvature, as measured by the Laplacian of the PINN loss.
arXiv Detail & Related papers (2022-10-14T15:01:32Z) - Equivariant Transduction through Invariant Alignment [71.45263447328374]
We introduce a novel group-equivariant architecture that incorporates a group-in hard alignment mechanism.
We find that our network's structure allows it to develop stronger equivariant properties than existing group-equivariant approaches.
We additionally find that it outperforms previous group-equivariant networks empirically on the SCAN task.
arXiv Detail & Related papers (2022-09-22T11:19:45Z) - Physics-Guided Problem Decomposition for Scaling Deep Learning of
High-dimensional Eigen-Solvers: The Case of Schr\"{o}dinger's Equation [8.80823317679047]
Deep neural networks (NNs) have been proposed as a viable alternative to traditional simulation-driven approaches for solving high-dimensional eigenvalue equations.
In this paper, we use physics knowledge to decompose the complex regression task of predicting the high-dimensional eigenvectors into simpler sub-tasks.
We demonstrate the efficacy of such physics-guided problem decomposition for the case of the Schr"odinger's Equation in Quantum Mechanics.
arXiv Detail & Related papers (2022-02-12T05:59:08Z) - Lorentz Group Equivariant Neural Network for Particle Physics [58.56031187968692]
We present a neural network architecture that is fully equivariant with respect to transformations under the Lorentz group.
For classification tasks in particle physics, we demonstrate that such an equivariant architecture leads to drastically simpler models that have relatively few learnable parameters.
arXiv Detail & Related papers (2020-06-08T17:54:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.