Safety design concepts for statistical machine learning components
toward accordance with functional safety standards
- URL: http://arxiv.org/abs/2008.01263v1
- Date: Tue, 4 Aug 2020 01:01:00 GMT
- Title: Safety design concepts for statistical machine learning components
toward accordance with functional safety standards
- Authors: Akihisa Morikawa and Yutaka Matsubara
- Abstract summary: In recent years, curial incidents and accidents have been reported due to misjudgment of statistical machine learning.
In this paper, we organize five kinds of technical safety concepts (TSCs) for components toward accordance with functional safety standards.
- Score: 0.38073142980732994
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, curial incidents and accidents have been reported due to
un-intended control caused by misjudgment of statistical machine learning
(SML), which include deep learning. The international functional safety
standards for Electric/Electronic/Programmable (E/E/P) systems have been widely
spread to improve safety. However, most of them do not recom-mended to use SML
in safety critical systems so far. In practical the new concepts and methods
are urgently required to enable SML to be safely used in safety critical
systems. In this paper, we organize five kinds of technical safety concepts
(TSCs) for SML components toward accordance with functional safety standards.
We discuss not only quantitative evaluation criteria, but also development
process based on XAI (eXplainable Artificial Intelligence) and Automotive SPICE
to improve explainability and reliability in development phase. Fi-nally, we
briefly compare the TSCs in cost and difficulty, and expect to en-courage
further discussion in many communities and domain.
Related papers
- Safety Monitoring of Machine Learning Perception Functions: a Survey [7.193217430660011]
New dependability challenges arise when Machine Learning predictions are used in safety-critical applications.
The use of fault tolerance mechanisms, such as safety monitors, is essential to ensure the safe behavior of the system.
This paper presents an extensive literature review on safety monitoring of perception functions using ML in a safety-critical context.
arXiv Detail & Related papers (2024-12-09T10:58:50Z) - Defining and Evaluating Physical Safety for Large Language Models [62.4971588282174]
Large Language Models (LLMs) are increasingly used to control robotic systems such as drones.
Their risks of causing physical threats and harm in real-world applications remain unexplored.
We classify the physical safety risks of drones into four categories: (1) human-targeted threats, (2) object-targeted threats, (3) infrastructure attacks, and (4) regulatory violations.
arXiv Detail & Related papers (2024-11-04T17:41:25Z) - Multimodal Situational Safety [73.63981779844916]
We present the first evaluation and analysis of a novel safety challenge termed Multimodal Situational Safety.
For an MLLM to respond safely, whether through language or action, it often needs to assess the safety implications of a language query within its corresponding visual context.
We develop the Multimodal Situational Safety benchmark (MSSBench) to assess the situational safety performance of current MLLMs.
arXiv Detail & Related papers (2024-10-08T16:16:07Z) - SCANS: Mitigating the Exaggerated Safety for LLMs via Safety-Conscious Activation Steering [56.92068213969036]
Safety alignment is indispensable for Large Language Models (LLMs) to defend threats from malicious instructions.
Recent researches reveal safety-aligned LLMs prone to reject benign queries due to the exaggerated safety issue.
We propose a Safety-Conscious Activation Steering (SCANS) method to mitigate the exaggerated safety concerns.
arXiv Detail & Related papers (2024-08-21T10:01:34Z) - SAFETY-J: Evaluating Safety with Critique [24.723999605458832]
We introduce SAFETY-J, a bilingual generative safety evaluator for English and Chinese with critique-based judgment.
We establish an automated meta-evaluation benchmark that objectively assesses the quality of critiques with minimal human intervention.
Our evaluations demonstrate that SAFETY-J provides more nuanced and accurate safety evaluations, thereby enhancing both critique quality and predictive reliability in complex content scenarios.
arXiv Detail & Related papers (2024-07-24T08:04:00Z) - Safety Analysis of Autonomous Railway Systems: An Introduction to the SACRED Methodology [2.47737926497181]
We introduce SACRED, a safety methodology for producing an initial safety case for autonomous systems.
The development of SACRED is motivated by the proposed GoA-4 light-rail system in Berlin.
arXiv Detail & Related papers (2024-03-18T11:12:19Z) - The Art of Defending: A Systematic Evaluation and Analysis of LLM
Defense Strategies on Safety and Over-Defensiveness [56.174255970895466]
Large Language Models (LLMs) play an increasingly pivotal role in natural language processing applications.
This paper presents Safety and Over-Defensiveness Evaluation (SODE) benchmark.
arXiv Detail & Related papers (2023-12-30T17:37:06Z) - Leveraging Traceability to Integrate Safety Analysis Artifacts into the
Software Development Process [51.42800587382228]
Safety assurance cases (SACs) can be challenging to maintain during system evolution.
We propose a solution that leverages software traceability to connect relevant system artifacts to safety analysis models.
We elicit design rationales for system changes to help safety stakeholders analyze the impact of system changes on safety.
arXiv Detail & Related papers (2023-07-14T16:03:27Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Reliability Assessment and Safety Arguments for Machine Learning
Components in Assuring Learning-Enabled Autonomous Systems [19.65793237440738]
We present an overall assurance framework for Learning-Enabled Systems (LES)
We then introduce a novel model-agnostic Reliability Assessment Model (RAM) for ML classifiers.
We discuss the model assumptions and the inherent challenges of assessing ML reliability uncovered by our RAM.
arXiv Detail & Related papers (2021-11-30T14:39:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.