Particle Transformer for Jet Tagging
- URL: http://arxiv.org/abs/2202.03772v3
- Date: Mon, 29 Jan 2024 15:22:51 GMT
- Title: Particle Transformer for Jet Tagging
- Authors: Huilin Qu, Congqiao Li, Sitian Qian
- Abstract summary: We present JetClass, a new comprehensive dataset for jet tagging.
The dataset consists of 100 M jets, about two orders of magnitude larger than existing public datasets.
We propose a new Transformer-based architecture for jet tagging, called Particle Transformer (ParT)
- Score: 4.604003661048267
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Jet tagging is a critical yet challenging classification task in particle
physics. While deep learning has transformed jet tagging and significantly
improved performance, the lack of a large-scale public dataset impedes further
enhancement. In this work, we present JetClass, a new comprehensive dataset for
jet tagging. The JetClass dataset consists of 100 M jets, about two orders of
magnitude larger than existing public datasets. A total of 10 types of jets are
simulated, including several types unexplored for tagging so far. Based on the
large dataset, we propose a new Transformer-based architecture for jet tagging,
called Particle Transformer (ParT). By incorporating pairwise particle
interactions in the attention mechanism, ParT achieves higher tagging
performance than a plain Transformer and surpasses the previous
state-of-the-art, ParticleNet, by a large margin. The pre-trained ParT models,
once fine-tuned, also substantially enhance the performance on two widely
adopted jet tagging benchmarks. The dataset, code and models are publicly
available at https://github.com/jet-universe/particle_transformer.
Related papers
- A multicategory jet image classification framework using deep neural network [0.9350546589421261]
Authors focus on jet category separability by particle and jet feature extraction, resulting in a computational efficient interpretable model for jet classification.
This work demonstrates that high dimensional datasets represented in separable latent spaces lead to simpler architectures for jet classification.
arXiv Detail & Related papers (2024-07-03T22:00:35Z) - Particle Multi-Axis Transformer for Jet Tagging [0.774199694856838]
In this article, we propose an idea of a new architecture, Particle Multi-Axis transformer (ParMAT)
ParMAT contains local and global spatial interactions within a single unit which improves its ability to handle various input lengths.
We trained our model on JETCLASS, a publicly available large dataset that contains 100M jets of 10 different classes of particles.
arXiv Detail & Related papers (2024-06-09T10:34:16Z) - OmniJet-$α$: The first cross-task foundation model for particle physics [0.0]
Foundation models are multi-dataset and multi-task machine learning methods that once pre-trained can be fine-tuned for a variety of downstream applications.
We report significant progress on this challenge on several fronts.
We demonstrate transfer learning between an unsupervised problem (jet generation) and a classic supervised task (jet tagging) with our new OmniJet-$alpha$ model.
arXiv Detail & Related papers (2024-03-08T19:00:01Z) - Flow Matching Beyond Kinematics: Generating Jets with Particle-ID and
Trajectory Displacement Information [0.0]
We introduce the first generative model trained on the JetClass dataset.
Our model generates jets at the constituent level, and it is a permutation-equivariant continuous normalizing flow (CNF) trained with the flow matching technique.
For the first time, we also introduce a generative model that goes beyond the kinematic features of jet constituents.
arXiv Detail & Related papers (2023-11-30T19:00:02Z) - JetSeg: Efficient Real-Time Semantic Segmentation Model for Low-Power
GPU-Embedded Systems [0.0]
We propose an efficient model for real-time semantic segmentation called JetSeg.
JetSeg consists of an encoder called JetNet, and an improved RegSeg decoder.
Our approach outperforms state-of-the-art real-time encoder-decoder models by reducing 46.70M parameters and 5.14% GFLOPs.
arXiv Detail & Related papers (2023-05-19T04:07:26Z) - Joint Spatial-Temporal and Appearance Modeling with Transformer for
Multiple Object Tracking [59.79252390626194]
We propose a novel solution named TransSTAM, which leverages Transformer to model both the appearance features of each object and the spatial-temporal relationships among objects.
The proposed method is evaluated on multiple public benchmarks including MOT16, MOT17, and MOT20, and it achieves a clear performance improvement in both IDF1 and HOTA.
arXiv Detail & Related papers (2022-05-31T01:19:18Z) - Autoencoders for unsupervised anomaly detection in high energy physics [105.54048699217668]
We study the tagging of top jet images in a background of QCD jet images.
We show that the standard autoencoder setup cannot be considered as a model-independent anomaly tagger.
We suggest improved performance measures for the task of model-independent anomaly detection.
arXiv Detail & Related papers (2021-04-19T05:06:57Z) - UniDrop: A Simple yet Effective Technique to Improve Transformer without
Extra Cost [110.67392881417777]
Transformer architecture achieves great success in abundant natural language processing tasks.
We find simple techniques such as dropout, can greatly boost model performance with a careful design.
Specifically, we propose an approach named UniDrop to unites three different dropout techniques.
arXiv Detail & Related papers (2021-04-11T07:43:19Z) - TransMOT: Spatial-Temporal Graph Transformer for Multiple Object
Tracking [74.82415271960315]
We propose a solution named TransMOT to efficiently model the spatial and temporal interactions among objects in a video.
TransMOT is not only more computationally efficient than the traditional Transformer, but it also achieves better tracking accuracy.
The proposed method is evaluated on multiple benchmark datasets including MOT15, MOT16, MOT17, and MOT20.
arXiv Detail & Related papers (2021-04-01T01:49:05Z) - ACRONYM: A Large-Scale Grasp Dataset Based on Simulation [64.37675024289857]
ACRONYM is a dataset for robot grasp planning based on physics simulation.
The dataset contains 17.7M parallel-jaw grasps, spanning 8872 objects from 262 different categories.
We show the value of this large and diverse dataset by using it to train two state-of-the-art learning-based grasp planning algorithms.
arXiv Detail & Related papers (2020-11-18T23:24:00Z) - The Cascade Transformer: an Application for Efficient Answer Sentence
Selection [116.09532365093659]
We introduce the Cascade Transformer, a technique to adapt transformer-based models into a cascade of rankers.
When compared to a state-of-the-art transformer model, our approach reduces computation by 37% with almost no impact on accuracy.
arXiv Detail & Related papers (2020-05-05T23:32:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.