Learning-based attacks in Cyber-Physical Systems: Exploration,
Detection, and Control Cost trade-offs
- URL: http://arxiv.org/abs/2011.10718v2
- Date: Thu, 20 May 2021 02:11:55 GMT
- Title: Learning-based attacks in Cyber-Physical Systems: Exploration,
Detection, and Control Cost trade-offs
- Authors: Anshuka Rangi, Mohammad Javad Khojasteh and Massimo Franceschetti
- Abstract summary: We study the problem of learning-based attacks in linear systems, where the communication channel between the controller and the plant can be hijacked by a malicious attacker.
We assume the attacker learns the dynamics of the system from observations, then overrides the controller's actuation signal.
We show that this bound is also order optimal, in the sense that if the attacker satisfies it, then there exists a learning algorithm with the given order expected deception time.
- Score: 9.453554184019108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of learning-based attacks in linear systems, where the
communication channel between the controller and the plant can be hijacked by a
malicious attacker. We assume the attacker learns the dynamics of the system
from observations, then overrides the controller's actuation signal, while
mimicking legitimate operation by providing fictitious sensor readings to the
controller. On the other hand, the controller is on a lookout to detect the
presence of the attacker and tries to enhance the detection performance by
carefully crafting its control signals. We study the trade-offs between the
information acquired by the attacker from observations, the detection
capabilities of the controller, and the control cost. Specifically, we provide
tight upper and lower bounds on the expected $\epsilon$-deception time, namely
the time required by the controller to make a decision regarding the presence
of an attacker with confidence at least $(1-\epsilon\log(1/\epsilon))$. We then
show a probabilistic lower bound on the time that must be spent by the attacker
learning the system, in order for the controller to have a given expected
$\epsilon$-deception time. We show that this bound is also order optimal, in
the sense that if the attacker satisfies it, then there exists a learning
algorithm with the given order expected deception time. Finally, we show a
lower bound on the expected energy expenditure required to guarantee detection
with confidence at least $1-\epsilon \log(1/\epsilon)$.
Related papers
- Federated Learning for Zero-Day Attack Detection in 5G and Beyond V2X Networks [9.86830550255822]
Connected and Automated Vehicles (CAVs) on top of 5G and Beyond networks (5GB) make them vulnerable to increasing vectors of security and privacy attacks.
We propose in this paper a novel detection mechanism that leverages the ability of the deep auto-encoder method to detect attacks relying only on the benign network traffic pattern.
Using federated learning, the proposed intrusion detection system can be trained with large and diverse benign network traffic, while preserving the CAVs privacy, and minimizing the communication overhead.
arXiv Detail & Related papers (2024-07-03T12:42:31Z) - Detecting and Mitigating System-Level Anomalies of Vision-Based Controllers [7.095058159492494]
Vision-based controllers can make erroneous predictions when faced with novel or out-of-distribution inputs.
In this work, we introduce a run-time anomaly monitor to detect and mitigate such closed-loop, system-level failures.
We validate the proposed approach on an autonomous aircraft taxiing system that uses a vision-based controller for taxiing.
arXiv Detail & Related papers (2023-09-23T20:33:38Z) - Model Extraction Attacks Against Reinforcement Learning Based
Controllers [9.273077240506016]
This paper focuses on the setting when a Deep Neural Network (DNN) controller is trained using Reinforcement Learning (RL) algorithms and is used to control a system.
In the first phase, also called the offline phase, the attacker uses side-channel information about the RL-reward function and the system dynamics to identify a set of candidate estimates of the unknown DNN.
In the second phase, also called the online phase, the attacker observes the behavior of the unknown DNN and uses these observations to shortlist the set of final policy estimates.
arXiv Detail & Related papers (2023-04-25T18:48:42Z) - Learning Decentralized Linear Quadratic Regulators with $\sqrt{T}$ Regret [1.529943343419486]
We propose an online learning algorithm that adaptively designs a decentralized linear quadratic regulator when the system model is unknown a priori.
We show that our controller enjoys an expected regret that scales as $sqrtT$ with the time horizon $T$ for the case of partially nested information pattern.
arXiv Detail & Related papers (2022-10-17T09:29:01Z) - Improving the Performance of Robust Control through Event-Triggered
Learning [74.57758188038375]
We propose an event-triggered learning algorithm that decides when to learn in the face of uncertainty in the LQR problem.
We demonstrate improved performance over a robust controller baseline in a numerical example.
arXiv Detail & Related papers (2022-07-28T17:36:37Z) - Illusory Attacks: Information-Theoretic Detectability Matters in Adversarial Attacks [76.35478518372692]
We introduce epsilon-illusory, a novel form of adversarial attack on sequential decision-makers.
Compared to existing attacks, we empirically find epsilon-illusory to be significantly harder to detect with automated methods.
Our findings suggest the need for better anomaly detectors, as well as effective hardware- and system-level defenses.
arXiv Detail & Related papers (2022-07-20T19:49:09Z) - Deep Q-learning: a robust control approach [4.125187280299247]
We formulate an uncertain linear time-invariant model by means of the neural tangent kernel to describe learning.
We show the instability of learning and analyze the agent's behavior in frequency-domain.
Numerical simulations in different OpenAI Gym environments suggest that the $mathcalH_infty$ controlled learning performs slightly better than Double deep Q-learning.
arXiv Detail & Related papers (2022-01-21T09:47:34Z) - Sample-Efficient Safety Assurances using Conformal Prediction [57.92013073974406]
Early warning systems can provide alerts when an unsafe situation is imminent.
To reliably improve safety, these warning systems should have a provable false negative rate.
We present a framework that combines a statistical inference technique known as conformal prediction with a simulator of robot/environment dynamics.
arXiv Detail & Related papers (2021-09-28T23:00:30Z) - No Need to Know Physics: Resilience of Process-based Model-free Anomaly
Detection for Industrial Control Systems [95.54151664013011]
We present a novel framework to generate adversarial spoofing signals that violate physical properties of the system.
We analyze four anomaly detectors published at top security conferences.
arXiv Detail & Related papers (2020-12-07T11:02:44Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z) - Improper Learning for Non-Stochastic Control [78.65807250350755]
We consider the problem of controlling a possibly unknown linear dynamical system with adversarial perturbations, adversarially chosen convex loss functions, and partially observed states.
Applying online descent to this parametrization yields a new controller which attains sublinear regret vs. a large class of closed-loop policies.
Our bounds are the first in the non-stochastic control setting that compete with emphall stabilizing linear dynamical controllers.
arXiv Detail & Related papers (2020-01-25T02:12:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.