Towards a Robust and Trustworthy Machine Learning System Development
- URL: http://arxiv.org/abs/2101.03042v1
- Date: Fri, 8 Jan 2021 14:43:58 GMT
- Title: Towards a Robust and Trustworthy Machine Learning System Development
- Authors: Pulei Xiong, Scott Buffett, Shahrear Iqbal, Philippe Lamontagne,
Mohammad Mamun, and Heather Molyneaux
- Abstract summary: We present our recent survey on the state-of-the-art ML trustworthiness and technologies from a security engineering perspective.
We then push our studies forward above and beyond a survey by describing a metamodel we created that represents the body of knowledge in a standard and visualized way for ML practitioners.
We propose future research directions motivated by our findings to advance the development of robust and trustworthy ML systems.
- Score: 0.09236074230806578
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine Learning (ML) technologies have been widely adopted in many mission
critical fields, such as cyber security, autonomous vehicle control,
healthcare, etc. to support intelligent decision-making. While ML has
demonstrated impressive performance over conventional methods in these
applications, concerns arose with respect to system resilience against
ML-specific security attacks and privacy breaches as well as the trust that
users have in these systems. In this article, firstly we present our recent
systematic and comprehensive survey on the state-of-the-art ML robustness and
trustworthiness technologies from a security engineering perspective, which
covers all aspects of secure ML system development including threat modeling,
common offensive and defensive technologies, privacy-preserving machine
learning, user trust in the context of machine learning, and empirical
evaluation for ML model robustness. Secondly, we then push our studies forward
above and beyond a survey by describing a metamodel we created that represents
the body of knowledge in a standard and visualized way for ML practitioners. We
further illustrate how to leverage the metamodel to guide a systematic threat
analysis and security design process in a context of generic ML system
development, which extends and scales up the classic process. Thirdly, we
propose future research directions motivated by our findings to advance the
development of robust and trustworthy ML systems. Our work differs from
existing surveys in this area in that, to the best of our knowledge, it is the
first of its kind of engineering effort to (i) explore the fundamental
principles and best practices to support robust and trustworthy ML system
development; and (ii) study the interplay of robustness and user trust in the
context of ML systems.
Related papers
- Towards Trustworthy Machine Learning in Production: An Overview of the Robustness in MLOps Approach [0.0]
In recent years, AI researchers and practitioners have introduced principles and guidelines to build systems that make reliable and trustworthy decisions.
In practice, a fundamental challenge arises when the system needs to be operationalized and deployed to evolve and operate in real-life environments continuously.
To address this challenge, Machine Learning Operations (MLOps) have emerged as a potential recipe for standardizing ML solutions in deployment.
arXiv Detail & Related papers (2024-10-28T09:34:08Z) - Empowering Autonomous Driving with Large Language Models: A Safety Perspective [82.90376711290808]
This paper explores the integration of Large Language Models (LLMs) into Autonomous Driving systems.
LLMs are intelligent decision-makers in behavioral planning, augmented with a safety verifier shield for contextual safety learning.
We present two key studies in a simulated environment: an adaptive LLM-conditioned Model Predictive Control (MPC) and an LLM-enabled interactive behavior planning scheme with a state machine.
arXiv Detail & Related papers (2023-11-28T03:13:09Z) - Vulnerability of Machine Learning Approaches Applied in IoT-based Smart Grid: A Review [51.31851488650698]
Machine learning (ML) sees an increasing prevalence of being used in the internet-of-things (IoT)-based smart grid.
adversarial distortion injected into the power signal will greatly affect the system's normal control and operation.
It is imperative to conduct vulnerability assessment for MLsgAPPs applied in the context of safety-critical power systems.
arXiv Detail & Related papers (2023-08-30T03:29:26Z) - Machine Learning with Confidential Computing: A Systematization of Knowledge [9.632031075287047]
Privacy and security challenges in Machine Learning (ML) have become increasingly severe, along with ML's pervasive development and the recent demonstration of large attack surfaces.
As a mature system-oriented approach, Confidential Computing has been utilized in both academia and industry to mitigate privacy and security issues in various ML scenarios.
We systematize the prior work on Confidential Computing-assisted ML techniques that provide i) confidentiality guarantees and ii) integrity assurances, and discuss their advanced features and drawbacks.
arXiv Detail & Related papers (2022-08-22T08:23:53Z) - SoK: Machine Learning Governance [16.36671448193025]
We develop the concept of ML governance to balance such benefits and risks.
We use identities to hold principals accountable for failures of ML systems.
We highlight the need for techniques that allow a model owner to manage the life cycle of their system.
arXiv Detail & Related papers (2021-09-20T17:56:22Z) - Multi Agent System for Machine Learning Under Uncertainty in Cyber
Physical Manufacturing System [78.60415450507706]
Recent advancements in predictive machine learning has led to its application in various use cases in manufacturing.
Most research focused on maximising predictive accuracy without addressing the uncertainty associated with it.
In this paper, we determine the sources of uncertainty in machine learning and establish the success criteria of a machine learning system to function well under uncertainty.
arXiv Detail & Related papers (2021-07-28T10:28:05Z) - Technology Readiness Levels for Machine Learning Systems [107.56979560568232]
Development and deployment of machine learning systems can be executed easily with modern tools, but the process is typically rushed and means-to-an-end.
We have developed a proven systems engineering approach for machine learning development and deployment.
Our "Machine Learning Technology Readiness Levels" framework defines a principled process to ensure robust, reliable, and responsible systems.
arXiv Detail & Related papers (2021-01-11T15:54:48Z) - Robust Machine Learning Systems: Challenges, Current Trends,
Perspectives, and the Road Ahead [24.60052335548398]
Machine Learning (ML) techniques have been rapidly adopted by smart Cyber-Physical Systems (CPS) and Internet-of-Things (IoT)
They are vulnerable to various security and reliability threats, at both hardware and software levels, that compromise their accuracy.
This paper summarizes the prominent vulnerabilities of modern ML systems, highlights successful defenses and mitigation techniques against these vulnerabilities.
arXiv Detail & Related papers (2021-01-04T20:06:56Z) - Dos and Don'ts of Machine Learning in Computer Security [74.1816306998445]
Despite great potential, machine learning in security is prone to subtle pitfalls that undermine its performance.
We identify common pitfalls in the design, implementation, and evaluation of learning-based security systems.
We propose actionable recommendations to support researchers in avoiding or mitigating the pitfalls where possible.
arXiv Detail & Related papers (2020-10-19T13:09:31Z) - Security and Machine Learning in the Real World [33.40597438876848]
We build on our experience evaluating the security of a machine learning software product deployed on a large scale to broaden the conversation to include a systems security view of vulnerabilities.
We propose a list of short-term mitigation suggestions that practitioners deploying machine learning modules can use to secure their systems.
arXiv Detail & Related papers (2020-07-13T16:57:12Z) - Technology Readiness Levels for AI & ML [79.22051549519989]
Development of machine learning systems can be executed easily with modern tools, but the process is typically rushed and means-to-an-end.
Engineering systems follow well-defined processes and testing standards to streamline development for high-quality, reliable results.
We propose a proven systems engineering approach for machine learning development and deployment.
arXiv Detail & Related papers (2020-06-21T17:14:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.