19 Parameters Is All You Need: Tiny Neural Networks for Particle Physics
- URL: http://arxiv.org/abs/2310.16121v3
- Date: Wed, 13 Dec 2023 20:56:06 GMT
- Title: 19 Parameters Is All You Need: Tiny Neural Networks for Particle Physics
- Authors: Alexander Bogatskiy, Timothy Hoffman, Jan T. Offermann
- Abstract summary: We present the potential of one recent Lorentz- and permutation-symmetric architecture, PELICAN, for low-latency neural network tasks.
We show its instances with as few as 19 trainable parameters that outperform generic architectures with tens of thousands of parameters when compared on the binary classification task of top quark jet tagging.
- Score: 52.42485649300583
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As particle accelerators increase their collision rates, and deep learning
solutions prove their viability, there is a growing need for lightweight and
fast neural network architectures for low-latency tasks such as triggering. We
examine the potential of one recent Lorentz- and permutation-symmetric
architecture, PELICAN, and present its instances with as few as 19 trainable
parameters that outperform generic architectures with tens of thousands of
parameters when compared on the binary classification task of top quark jet
tagging.
Related papers
- Explainable Equivariant Neural Networks for Particle Physics: PELICAN [51.02649432050852]
PELICAN is a novel permutation equivariant and Lorentz invariant aggregator network.
We present a study of the PELICAN algorithm architecture in the context of both tagging (classification) and reconstructing (regression) Lorentz-boosted top quarks.
We extend the application of PELICAN to the tasks of identifying quark-initiated vs.gluon-initiated jets, and a multi-class identification across five separate target categories of jets.
arXiv Detail & Related papers (2023-07-31T09:08:40Z) - Parameter-efficient Tuning of Large-scale Multimodal Foundation Model [68.24510810095802]
We propose A graceful prompt framework for cross-modal transfer (Aurora) to overcome these challenges.
Considering the redundancy in existing architectures, we first utilize the mode approximation to generate 0.1M trainable parameters to implement the multimodal prompt tuning.
A thorough evaluation on six cross-modal benchmarks shows that it not only outperforms the state-of-the-art but even outperforms the full fine-tuning approach.
arXiv Detail & Related papers (2023-05-15T06:40:56Z) - Quantum HyperNetworks: Training Binary Neural Networks in Quantum
Superposition [16.1356415877484]
We introduce quantum hypernetworks as a mechanism to train binary neural networks on quantum computers.
We show that our approach effectively finds optimal parameters, hyperparameters and architectural choices with high probability on classification problems.
Our unified approach provides an immense scope for other applications in the field of machine learning.
arXiv Detail & Related papers (2023-01-19T20:06:48Z) - PELICAN: Permutation Equivariant and Lorentz Invariant or Covariant
Aggregator Network for Particle Physics [64.5726087590283]
We present a machine learning architecture that uses a set of inputs maximally reduced with respect to the full 6-dimensional Lorentz symmetry.
We show that the resulting network outperforms all existing competitors despite much lower model complexity.
arXiv Detail & Related papers (2022-11-01T13:36:50Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Semi-Equivariant GNN Architectures for Jet Tagging [1.6626046865692057]
We present the novel architecture VecNet that combines symmetry-respecting and unconstrained operations to study and tune the degree of physics-informed GNNs.
We find that a generalized architecture such as ours can deliver optimal performance in resource-constrained applications.
arXiv Detail & Related papers (2022-02-14T18:57:12Z) - Lorentz Group Equivariant Neural Network for Particle Physics [58.56031187968692]
We present a neural network architecture that is fully equivariant with respect to transformations under the Lorentz group.
For classification tasks in particle physics, we demonstrate that such an equivariant architecture leads to drastically simpler models that have relatively few learnable parameters.
arXiv Detail & Related papers (2020-06-08T17:54:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.