Efficient Dynamic Shielding for Parametric Safety Specifications
- URL: http://arxiv.org/abs/2505.22104v1
- Date: Wed, 28 May 2025 08:30:03 GMT
- Title: Efficient Dynamic Shielding for Parametric Safety Specifications
- Authors: Davide Corsi, Kaushik Mallik, Andoni Rodriguez, Cesar Sanchez,
- Abstract summary: Shielding is a runtime safety enforcement tool that needs to monitor and intervene the AI controller's actions if safety could be compromised otherwise.<n>We introduce dynamic shields for parametric safety specifications, which are succinctly represented sets of all possible safety specifications that may be encountered at runtime.<n>In our experiments, the dynamic shields took a few minutes for their offline design, and took between a fraction of a second and a few seconds for online adaptation at each step, whereas the brute-force online recomputation approach was up to 5 times slower.
- Score: 2.1829548755022423
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Shielding has emerged as a promising approach for ensuring safety of AI-controlled autonomous systems. The algorithmic goal is to compute a shield, which is a runtime safety enforcement tool that needs to monitor and intervene the AI controller's actions if safety could be compromised otherwise. Traditional shields are designed statically for a specific safety requirement. Therefore, if the safety requirement changes at runtime due to changing operating conditions, the shield needs to be recomputed from scratch, causing delays that could be fatal. We introduce dynamic shields for parametric safety specifications, which are succinctly represented sets of all possible safety specifications that may be encountered at runtime. Our dynamic shields are statically designed for a given safety parameter set, and are able to dynamically adapt as the true safety specification (permissible by the parameters) is revealed at runtime. The main algorithmic novelty lies in the dynamic adaptation procedure, which is a simple and fast algorithm that utilizes known features of standard safety shields, like maximal permissiveness. We report experimental results for a robot navigation problem in unknown territories, where the safety specification evolves as new obstacles are discovered at runtime. In our experiments, the dynamic shields took a few minutes for their offline design, and took between a fraction of a second and a few seconds for online adaptation at each step, whereas the brute-force online recomputation approach was up to 5 times slower.
Related papers
- Shape it Up! Restoring LLM Safety during Finetuning [66.46166656543761]
Finetuning large language models (LLMs) enables user-specific customization but introduces critical safety risks.<n>We propose dynamic safety shaping (DSS), a framework that uses fine-grained safety signals to reinforce learning from safe segments of a response while suppressing unsafe content.<n>We present STAR-DSS, guided by STAR scores, that robustly mitigates finetuning risks and delivers substantial safety improvements across diverse threats, datasets, and model families.
arXiv Detail & Related papers (2025-05-22T18:05:16Z) - Runtime Safety through Adaptive Shielding: From Hidden Parameter Inference to Provable Guarantees [17.670635109868854]
Variations in hidden parameters, such as a robot's mass distribution or friction, pose safety risks during execution.<n>We develop a runtime shielding mechanism for reinforcement learning.<n>We prove that the proposed mechanism satisfies probabilistic safety guarantees and yields optimal policies.
arXiv Detail & Related papers (2025-05-20T23:45:45Z) - One Filter to Deploy Them All: Robust Safety for Quadrupedal Navigation in Unknown Environments [10.103090762559708]
We propose an observation-conditioned reachability-based (OCR) safety-filter framework.<n>Our key idea is to use an OCR value network (OCR-VN) that predicts the optimal control-theoretic safety value function for new failure regions.<n>We demonstrate that the proposed framework can automatically safeguard a wide range of hierarchical quadruped controllers.
arXiv Detail & Related papers (2024-12-13T09:21:02Z) - Compositional Shielding and Reinforcement Learning for Multi-Agent Systems [1.124958340749622]
Deep reinforcement learning has emerged as a powerful tool for obtaining high-performance policies.
One promising paradigm to guarantee safety is a shield, which shields a policy from making unsafe actions.
In this work, we propose a novel approach for multi-agent shielding.
arXiv Detail & Related papers (2024-10-14T12:52:48Z) - Realizable Continuous-Space Shields for Safe Reinforcement Learning [13.728961635717134]
We present the first shielding approach specifically designed to ensure the satisfaction of safety requirements in continuous state and action spaces.<n>Our method builds upon realizability, an essential property that confirms the shield will always be able to generate a safe action for any state in the environment.
arXiv Detail & Related papers (2024-10-02T21:08:11Z) - Safety through Permissibility: Shield Construction for Fast and Safe Reinforcement Learning [57.84059344739159]
"Shielding" is a popular technique to enforce safety inReinforcement Learning (RL)
We propose a new permissibility-based framework to deal with safety and shield construction.
arXiv Detail & Related papers (2024-05-29T18:00:21Z) - Safety Shielding under Delayed Observation [59.86192283565134]
Shields are correct-by-construction enforcers that guarantee safe execution.
Shields should pick the safe corrective actions in such a way that future interferences are most likely minimized.
We present the first integration of shields in a realistic driving simulator.
arXiv Detail & Related papers (2023-07-05T10:06:10Z) - Approximate Shielding of Atari Agents for Safe Exploration [83.55437924143615]
We propose a principled algorithm for safe exploration based on the concept of shielding.
We present preliminary results that show our approximate shielding algorithm effectively reduces the rate of safety violations.
arXiv Detail & Related papers (2023-04-21T16:19:54Z) - Online Shielding for Reinforcement Learning [59.86192283565134]
We propose an approach for online safety shielding of RL agents.
During runtime, the shield analyses the safety of each available action.
Based on this probability and a given threshold, the shield decides whether to block an action from the agent.
arXiv Detail & Related papers (2022-12-04T16:00:29Z) - BarrierNet: A Safety-Guaranteed Layer for Neural Networks [50.86816322277293]
BarrierNet allows the safety constraints of a neural controller be adaptable to changing environments.
We evaluate them on a series of control problems such as traffic merging and robot navigations in 2D and 3D space.
arXiv Detail & Related papers (2021-11-22T15:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.