LEAN-DMKDE: Quantum Latent Density Estimation for Anomaly Detection
- URL: http://arxiv.org/abs/2211.08525v1
- Date: Tue, 15 Nov 2022 21:51:42 GMT
- Title: LEAN-DMKDE: Quantum Latent Density Estimation for Anomaly Detection
- Authors: Joseph Gallego-Mejia, Oscar Bustos-Brinez, Fabio A. Gonz\'alez
- Abstract summary: The method combines an autoencoder, for learning a low-dimensional representation of the data, with a density-estimation model.
The method predicts a degree of normality for new samples based on the estimated density.
The experimental results show that the method performs on par with or outperforms other state-of-the-art methods.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents an anomaly detection model that combines the strong
statistical foundation of density-estimation-based anomaly detection methods
with the representation-learning ability of deep-learning models. The method
combines an autoencoder, for learning a low-dimensional representation of the
data, with a density-estimation model based on random Fourier features and
density matrices in an end-to-end architecture that can be trained using
gradient-based optimization techniques. The method predicts a degree of
normality for new samples based on the estimated density. A systematic
experimental evaluation was performed on different benchmark datasets. The
experimental results show that the method performs on par with or outperforms
other state-of-the-art methods.
Related papers
- Supervised Score-Based Modeling by Gradient Boosting [49.556736252628745]
We propose a Supervised Score-based Model (SSM) which can be viewed as a gradient boosting algorithm combining score matching.
We provide a theoretical analysis of learning and sampling for SSM to balance inference time and prediction accuracy.
Our model outperforms existing models in both accuracy and inference time.
arXiv Detail & Related papers (2024-11-02T07:06:53Z) - Latent Anomaly Detection Through Density Matrices [3.843839245375552]
This paper introduces a novel anomaly detection framework that combines the robust statistical principles of density-estimation-based anomaly detection methods with the representation-learning capabilities of deep learning models.
The method originated from this framework is presented in two different versions: a shallow approach and a deep approach that integrates an autoencoder to learn a low-dimensional representation of the data.
arXiv Detail & Related papers (2024-08-14T15:44:51Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - A Geometric Perspective on Diffusion Models [57.27857591493788]
We inspect the ODE-based sampling of a popular variance-exploding SDE.
We establish a theoretical relationship between the optimal ODE-based sampling and the classic mean-shift (mode-seeking) algorithm.
arXiv Detail & Related papers (2023-05-31T15:33:16Z) - Efficient Training of Energy-Based Models Using Jarzynski Equality [13.636994997309307]
Energy-based models (EBMs) are generative models inspired by statistical physics.
The computation of its gradient with respect to the model parameters requires sampling the model distribution.
Here we show how results for nonequilibrium thermodynamics based on Jarzynski equality can be used to perform this computation efficiently.
arXiv Detail & Related papers (2023-05-30T21:07:52Z) - AD-DMKDE: Anomaly Detection through Density Matrices and Fourier
Features [0.0]
The method can be seen as an efficient approximation of Kernel Density Estimation (KDE)
A systematic comparison of the proposed method with eleven state-of-the-art anomaly detection methods on various data sets is presented.
arXiv Detail & Related papers (2022-10-26T15:43:16Z) - Quantum Adaptive Fourier Features for Neural Density Estimation [0.0]
This paper presents a method for neural density estimation that can be seen as a type of kernel density estimation.
The method is based on density matrices, a formalism used in quantum mechanics, and adaptive Fourier features.
The method was evaluated in different synthetic and real datasets, and its performance compared against state-of-the-art neural density estimation methods.
arXiv Detail & Related papers (2022-08-01T01:39:11Z) - Efficient training of lightweight neural networks using Online
Self-Acquired Knowledge Distillation [51.66271681532262]
Online Self-Acquired Knowledge Distillation (OSAKD) is proposed, aiming to improve the performance of any deep neural model in an online manner.
We utilize k-nn non-parametric density estimation technique for estimating the unknown probability distributions of the data samples in the output feature space.
arXiv Detail & Related papers (2021-08-26T14:01:04Z) - Meta-learning One-class Classifiers with Eigenvalue Solvers for
Supervised Anomaly Detection [55.888835686183995]
We propose a neural network-based meta-learning method for supervised anomaly detection.
We experimentally demonstrate that the proposed method achieves better performance than existing anomaly detection and few-shot learning methods.
arXiv Detail & Related papers (2021-03-01T01:43:04Z) - Interpolation Technique to Speed Up Gradients Propagation in Neural ODEs [71.26657499537366]
We propose a simple literature-based method for the efficient approximation of gradients in neural ODE models.
We compare it with the reverse dynamic method to train neural ODEs on classification, density estimation, and inference approximation tasks.
arXiv Detail & Related papers (2020-03-11T13:15:57Z) - Learning Generative Models using Denoising Density Estimators [29.068491722778827]
We introduce a new generative model based on denoising density estimators (DDEs)
Our main contribution is a novel technique to obtain generative models by minimizing the KL-divergence directly.
Experimental results demonstrate substantial improvement in density estimation and competitive performance in generative model training.
arXiv Detail & Related papers (2020-01-08T20:30:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.