Optimal Sub-sampling to Boost Power of Kernel Sequential Change-point
Detection
- URL: http://arxiv.org/abs/2210.15060v1
- Date: Wed, 26 Oct 2022 22:05:29 GMT
- Title: Optimal Sub-sampling to Boost Power of Kernel Sequential Change-point
Detection
- Authors: Song Wei, Chaofan Huang, Yao Xie
- Abstract summary: We present a novel scheme to boost detection power for kernel maximum mean discrepancy based sequential change-point detection procedures.
Our proposed scheme features an optimal sub-sampling of the history data before the detection procedure, in order to tackle the power loss incurred by the random sub-sample from the enormous history data.
- Score: 10.969806056391004
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We present a novel scheme to boost detection power for kernel maximum mean
discrepancy based sequential change-point detection procedures. Our proposed
scheme features an optimal sub-sampling of the history data before the
detection procedure, in order to tackle the power loss incurred by the random
sub-sample from the enormous history data. We apply our proposed scheme to both
Scan $B$ and Kernel Cumulative Sum (CUSUM) procedures, and improved performance
is observed from extensive numerical experiments.
Related papers
- Enhancing Gaussian Process Surrogates for Optimization and Posterior Approximation via Random Exploration [2.984929040246293]
novel noise-free Bayesian optimization strategies that rely on a random exploration step to enhance the accuracy of Gaussian process surrogate models.
New algorithms retain the ease of implementation of the classical GP-UCB, but an additional exploration step facilitates their convergence.
arXiv Detail & Related papers (2024-01-30T14:16:06Z) - Optimal Budgeted Rejection Sampling for Generative Models [54.050498411883495]
Rejection sampling methods have been proposed to improve the performance of discriminator-based generative models.
We first propose an Optimal Budgeted Rejection Sampling scheme that is provably optimal.
Second, we propose an end-to-end method that incorporates the sampling scheme into the training procedure to further enhance the model's overall performance.
arXiv Detail & Related papers (2023-11-01T11:52:41Z) - Conditional Denoising Diffusion for Sequential Recommendation [62.127862728308045]
Two prominent generative models, Generative Adversarial Networks (GANs) and Variational AutoEncoders (VAEs)
GANs suffer from unstable optimization, while VAEs are prone to posterior collapse and over-smoothed generations.
We present a conditional denoising diffusion model, which includes a sequence encoder, a cross-attentive denoising decoder, and a step-wise diffuser.
arXiv Detail & Related papers (2023-04-22T15:32:59Z) - Online Kernel CUSUM for Change-Point Detection [12.383181837411469]
We present a computationally efficient online kernel Cumulative Sum (CUSUM) method for change-point detection.
Our approach exhibits increased sensitivity to small changes compared to existing kernel-based change-point detection methods.
arXiv Detail & Related papers (2022-11-28T05:08:30Z) - Large-Scale Sequential Learning for Recommender and Engineering Systems [91.3755431537592]
In this thesis, we focus on the design of an automatic algorithms that provide personalized ranking by adapting to the current conditions.
For the former, we propose novel algorithm called SAROS that take into account both kinds of feedback for learning over the sequence of interactions.
The proposed idea of taking into account the neighbour lines shows statistically significant results in comparison with the initial approach for faults detection in power grid.
arXiv Detail & Related papers (2022-05-13T21:09:41Z) - Efficient Neural Network Analysis with Sum-of-Infeasibilities [64.31536828511021]
Inspired by sum-of-infeasibilities methods in convex optimization, we propose a novel procedure for analyzing verification queries on networks with extensive branching functions.
An extension to a canonical case-analysis-based complete search procedure can be achieved by replacing the convex procedure executed at each search state with DeepSoI.
arXiv Detail & Related papers (2022-03-19T15:05:09Z) - SelectAugment: Hierarchical Deterministic Sample Selection for Data
Augmentation [72.58308581812149]
We propose an effective approach, dubbed SelectAugment, to select samples to be augmented in a deterministic and online manner.
Specifically, in each batch, we first determine the augmentation ratio, and then decide whether to augment each training sample under this ratio.
In this way, the negative effects of the randomness in selecting samples to augment can be effectively alleviated and the effectiveness of DA is improved.
arXiv Detail & Related papers (2021-12-06T08:38:38Z) - Unsupervised learning of disentangled representations in deep restricted
kernel machines with orthogonality constraints [15.296955630621566]
Constr-DRKM is a deep kernel method for the unsupervised learning of disentangled data representations.
We quantitatively evaluate the proposed method's effectiveness in disentangled feature learning.
arXiv Detail & Related papers (2020-11-25T11:40:10Z) - Bayesian Coresets: Revisiting the Nonconvex Optimization Perspective [30.963638533636352]
We propose and analyze a novel algorithm for coreset selection.
We provide explicit convergence rate guarantees and present an empirical evaluation on a variety of benchmark datasets.
arXiv Detail & Related papers (2020-07-01T19:34:59Z) - Robust Sampling in Deep Learning [62.997667081978825]
Deep learning requires regularization mechanisms to reduce overfitting and improve generalization.
We address this problem by a new regularization method based on distributional robust optimization.
During the training, the selection of samples is done according to their accuracy in such a way that the worst performed samples are the ones that contribute the most in the optimization.
arXiv Detail & Related papers (2020-06-04T09:46:52Z) - Sequential Adversarial Anomaly Detection for One-Class Event Data [18.577418448786634]
We consider the sequential anomaly detection problem in the one-class setting when only the anomalous sequences are available.
We propose an adversarial sequential detector by solving a minimax problem to find an optimal detector against the worst-case sequences from a generator.
arXiv Detail & Related papers (2019-10-21T06:12:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.