High-dimensional and Permutation Invariant Anomaly Detection
- URL: http://arxiv.org/abs/2306.03933v5
- Date: Wed, 7 Feb 2024 18:16:43 GMT
- Title: High-dimensional and Permutation Invariant Anomaly Detection
- Authors: Vinicius Mikuni, Benjamin Nachman
- Abstract summary: We introduce a permutation-invariant density estimator for particle physics data based on diffusion models.
We demonstrate the efficacy of our methodology by utilizing the learned density as a permutation-invariant anomaly detection score.
- Score: 0.1450405446885067
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Methods for anomaly detection of new physics processes are often limited to
low-dimensional spaces due to the difficulty of learning high-dimensional
probability densities. Particularly at the constituent level, incorporating
desirable properties such as permutation invariance and variable-length inputs
becomes difficult within popular density estimation methods. In this work, we
introduce a permutation-invariant density estimator for particle physics data
based on diffusion models, specifically designed to handle variable-length
inputs. We demonstrate the efficacy of our methodology by utilizing the learned
density as a permutation-invariant anomaly detection score, effectively
identifying jets with low likelihood under the background-only hypothesis. To
validate our density estimation method, we investigate the ratio of learned
densities and compare to those obtained by a supervised classification
algorithm.
Related papers
- Stochastic full waveform inversion with deep generative prior for uncertainty quantification [0.0]
Full Waveform Inversion (FWI) involves solving a nonlinear and often non-unique inverse problem.
FWI presents challenges such as local minima trapping and inadequate handling of inherent uncertainties.
We propose leveraging deep generative models as the prior distribution of geophysical parameters for Bayesian inversion.
arXiv Detail & Related papers (2024-06-07T11:44:50Z) - Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning [50.84938730450622]
We propose a trajectory-based method TV score, which uses trajectory volatility for OOD detection in mathematical reasoning.
Our method outperforms all traditional algorithms on GLMs under mathematical reasoning scenarios.
Our method can be extended to more applications with high-density features in output spaces, such as multiple-choice questions.
arXiv Detail & Related papers (2024-05-22T22:22:25Z) - TERM Model: Tensor Ring Mixture Model for Density Estimation [48.622060998018206]
In this paper, we take tensor ring decomposition for density estimator, which significantly reduces the number of permutation candidates.
A mixture model that incorporates multiple permutation candidates with adaptive weights is further designed, resulting in increased expressive flexibility.
This approach acknowledges that suboptimal permutations can offer distinctive information besides that of optimal permutations.
arXiv Detail & Related papers (2023-12-13T11:39:56Z) - Sobolev Space Regularised Pre Density Models [51.558848491038916]
We propose a new approach to non-parametric density estimation that is based on regularizing a Sobolev norm of the density.
This method is statistically consistent, and makes the inductive validation model clear and consistent.
arXiv Detail & Related papers (2023-07-25T18:47:53Z) - Anomaly Detection with Variance Stabilized Density Estimation [49.46356430493534]
We present a variance-stabilized density estimation problem for maximizing the likelihood of the observed samples.
To obtain a reliable anomaly detector, we introduce a spectral ensemble of autoregressive models for learning the variance-stabilized distribution.
We have conducted an extensive benchmark with 52 datasets, demonstrating that our method leads to state-of-the-art results.
arXiv Detail & Related papers (2023-06-01T11:52:58Z) - LEAN-DMKDE: Quantum Latent Density Estimation for Anomaly Detection [0.0]
The method combines an autoencoder, for learning a low-dimensional representation of the data, with a density-estimation model.
The method predicts a degree of normality for new samples based on the estimated density.
The experimental results show that the method performs on par with or outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2022-11-15T21:51:42Z) - Density-Based Clustering with Kernel Diffusion [59.4179549482505]
A naive density corresponding to the indicator function of a unit $d$-dimensional Euclidean ball is commonly used in density-based clustering algorithms.
We propose a new kernel diffusion density function, which is adaptive to data of varying local distributional characteristics and smoothness.
arXiv Detail & Related papers (2021-10-11T09:00:33Z) - Featurized Density Ratio Estimation [82.40706152910292]
In our work, we propose to leverage an invertible generative model to map the two distributions into a common feature space prior to estimation.
This featurization brings the densities closer together in latent space, sidestepping pathological scenarios where the learned density ratios in input space can be arbitrarily inaccurate.
At the same time, the invertibility of our feature map guarantees that the ratios computed in feature space are equivalent to those in input space.
arXiv Detail & Related papers (2021-07-05T18:30:26Z) - Annealed Stein Variational Gradient Descent [4.020523898765405]
Stein variational gradient descent has gained attention in the approximate literature inference for its flexibility and accuracy.
We empirically explore the ability of this method to sample from multi-modal distributions and focus on two important issues: (i) the inability of the particles to escape from local modes and (ii) the inefficacy in reproducing the density of the different regions.
arXiv Detail & Related papers (2021-01-24T22:18:30Z) - Low-rank Characteristic Tensor Density Estimation Part I: Foundations [38.05393186002834]
We propose a novel approach that builds upon tensor factorization tools.
In order to circumvent the curse of dimensionality, we introduce a low-rank model of this characteristic tensor.
We demonstrate the very promising performance of the proposed method using several measured datasets.
arXiv Detail & Related papers (2020-08-27T18:06:19Z) - Efficient sampling generation from explicit densities via Normalizing
Flows [0.0]
We will present a method based on normalizing flows, proposing a solution for the common problem of exploding reverse Kullback-Leibler divergence.
The performance of the method will be demonstrated using a multi-mode complex density function.
arXiv Detail & Related papers (2020-03-23T12:03:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.