Asymptotically Optimal Search for a Change Point Anomaly under a Composite Hypothesis Model
- URL: http://arxiv.org/abs/2412.19392v1
- Date: Fri, 27 Dec 2024 00:44:34 GMT
- Title: Asymptotically Optimal Search for a Change Point Anomaly under a Composite Hypothesis Model
- Authors: Liad Lea Didi, Tomer Gafni, Kobi Cohen,
- Abstract summary: We address the problem of searching for a change point in an anomalous process among a finite set of M processes.
Our goal is to design a sequential search strategy that minimizes the Bayes risk by balancing sample complexity and detection accuracy.
- Score: 10.514231683620517
- License:
- Abstract: We address the problem of searching for a change point in an anomalous process among a finite set of M processes. Specifically, we address a composite hypothesis model in which each process generates measurements following a common distribution with an unknown parameter (vector). This parameter belongs to either a normal or abnormal space depending on the current state of the process. Before the change point, all processes, including the anomalous one, are in a normal state; after the change point, the anomalous process transitions to an abnormal state. Our goal is to design a sequential search strategy that minimizes the Bayes risk by balancing sample complexity and detection accuracy. We propose a deterministic search algorithm with the following notable properties. First, we analytically demonstrate that when the distributions of both normal and abnormal processes are unknown, the algorithm is asymptotically optimal in minimizing the Bayes risk as the error probability approaches zero. In the second setting, where the parameter under the null hypothesis is known, the algorithm achieves asymptotic optimality with improved detection time based on the true normal state. Simulation results are presented to validate the theoretical findings.
Related papers
- Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Optimal Algorithms for the Inhomogeneous Spiked Wigner Model [89.1371983413931]
We derive an approximate message-passing algorithm (AMP) for the inhomogeneous problem.
We identify in particular the existence of a statistical-to-computational gap where known algorithms require a signal-to-noise ratio bigger than the information-theoretic threshold to perform better than random.
arXiv Detail & Related papers (2023-02-13T19:57:17Z) - Stochastic Approximation with Decision-Dependent Distributions: Asymptotic Normality and Optimality [8.771678221101368]
We analyze an approximation for decision-dependent problems, wherein the data distribution used by the algorithm evolves along the iterate sequence.
We show that under mild assumptions, the deviation between the iterate of the algorithm and its solution isally normal.
We also show that the performance of the algorithm with averaging is locally minimax optimal.
arXiv Detail & Related papers (2022-07-09T01:44:17Z) - Infinite-Dimensional Sparse Learning in Linear System Identification [0.2867517731896504]
This paper proposes an infinite-dimensional sparse learning algorithm based on atomic norm regularization.
The difficulty in solving the problem lies in the fact that there are an infinite number of possible atomic models.
arXiv Detail & Related papers (2022-03-28T13:18:48Z) - Sharp global convergence guarantees for iterative nonconvex
optimization: A Gaussian process perspective [30.524043513721168]
We develop a general recipe for analyzing the convergence of iterative algorithms for a class of regression models.
deterministicly, we accurately capture both the convergence rate of the algorithm and the eventual error floor in the finite-sample regime.
We show sharp convergence rates for both higher-order algorithms based on alternating updates and first-order algorithms based on subgradient subgradients.
arXiv Detail & Related papers (2021-09-20T21:48:19Z) - Quickest change detection with unknown parameters: Constant complexity
and near optimality [0.0]
We consider the quickest change detection problem where both the parameters of pre- and post- change distributions are unknown.
We derive a new scalable approximate algorithm with near optimal performance that runsin$mathcalO(1)$, adapted to even more complex Markovian settings.
arXiv Detail & Related papers (2021-06-09T13:29:26Z) - Anomaly Detection via Controlled Sensing and Deep Active Inference [43.07302992747749]
In this paper, we address the anomaly detection problem where the objective is to find the anomalous processes among a given set of processes.
We develop a sequential selection algorithm that decides which processes to be probed at every instant to detect the anomalies.
Our algorithm is based on active inference which is a general framework to make sequential decisions in order to maximize the notion of free energy.
arXiv Detail & Related papers (2021-05-12T17:54:02Z) - Optimal Sequential Detection of Signals with Unknown Appearance and
Disappearance Points in Time [64.26593350748401]
The paper addresses a sequential changepoint detection problem, assuming that the duration of change may be finite and unknown.
We focus on a reliable maximin change detection criterion of maximizing the minimal probability of detection in a given time (or space) window.
The FMA algorithm is applied to detecting faint streaks of satellites in optical images.
arXiv Detail & Related papers (2021-02-02T04:58:57Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - Active Model Estimation in Markov Decision Processes [108.46146218973189]
We study the problem of efficient exploration in order to learn an accurate model of an environment, modeled as a Markov decision process (MDP)
We show that our Markov-based algorithm outperforms both our original algorithm and the maximum entropy algorithm in the small sample regime.
arXiv Detail & Related papers (2020-03-06T16:17:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.