Closing the AI Accountability Gap: Defining an End-to-End Framework for
Internal Algorithmic Auditing
- URL: http://arxiv.org/abs/2001.00973v1
- Date: Fri, 3 Jan 2020 20:19:04 GMT
- Title: Closing the AI Accountability Gap: Defining an End-to-End Framework for
Internal Algorithmic Auditing
- Authors: Inioluwa Deborah Raji, Andrew Smart, Rebecca N. White, Margaret
Mitchell, Timnit Gebru, Ben Hutchinson, Jamila Smith-Loud, Daniel Theron,
Parker Barnes
- Abstract summary: We introduce a framework for algorithmic auditing that supports artificial intelligence system development end-to-end.
The proposed auditing framework is intended to close the accountability gap in the development and deployment of large-scale artificial intelligence systems.
- Score: 8.155332346712424
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Rising concern for the societal implications of artificial intelligence
systems has inspired a wave of academic and journalistic literature in which
deployed systems are audited for harm by investigators from outside the
organizations deploying the algorithms. However, it remains challenging for
practitioners to identify the harmful repercussions of their own systems prior
to deployment, and, once deployed, emergent issues can become difficult or
impossible to trace back to their source. In this paper, we introduce a
framework for algorithmic auditing that supports artificial intelligence system
development end-to-end, to be applied throughout the internal organization
development lifecycle. Each stage of the audit yields a set of documents that
together form an overall audit report, drawing on an organization's values or
principles to assess the fit of decisions made throughout the process. The
proposed auditing framework is intended to contribute to closing the
accountability gap in the development and deployment of large-scale artificial
intelligence systems by embedding a robust process to ensure audit integrity.
Related papers
- From Transparency to Accountability and Back: A Discussion of Access and Evidence in AI Auditing [1.196505602609637]
Audits can take many forms, including pre-deployment risk assessments, ongoing monitoring, and compliance testing.
There are many operational challenges to AI auditing that complicate its implementation.
We argue that auditing can be cast as a natural hypothesis test, draw parallels hypothesis testing and legal procedure, and argue that this framing provides clear and interpretable guidance on audit implementation.
arXiv Detail & Related papers (2024-10-07T06:15:46Z) - Trustworthiness in Retrieval-Augmented Generation Systems: A Survey [59.26328612791924]
Retrieval-Augmented Generation (RAG) has quickly grown into a pivotal paradigm in the development of Large Language Models (LLMs)
We propose a unified framework that assesses the trustworthiness of RAG systems across six key dimensions: factuality, robustness, fairness, transparency, accountability, and privacy.
arXiv Detail & Related papers (2024-09-16T09:06:44Z) - A Blueprint for Auditing Generative AI [0.9999629695552196]
generative AI systems display emergent capabilities and are adaptable to a wide range of downstream tasks.
Existing auditing procedures fail to address the governance challenges posed by generative AI systems.
We propose a three-layered approach, whereby governance audits of technology providers that design and disseminate generative AI systems, model audits of generative AI systems after pre-training but prior to their release, and application audits of applications based on top of generative AI systems.
arXiv Detail & Related papers (2024-07-07T11:56:54Z) - Towards Responsible AI in Banking: Addressing Bias for Fair
Decision-Making [69.44075077934914]
"Responsible AI" emphasizes the critical nature of addressing biases within the development of a corporate culture.
This thesis is structured around three fundamental pillars: understanding bias, mitigating bias, and accounting for bias.
In line with open-source principles, we have released Bias On Demand and FairView as accessible Python packages.
arXiv Detail & Related papers (2024-01-13T14:07:09Z) - Accountability in Offline Reinforcement Learning: Explaining Decisions
with a Corpus of Examples [70.84093873437425]
This paper introduces the Accountable Offline Controller (AOC) that employs the offline dataset as the Decision Corpus.
AOC operates effectively in low-data scenarios, can be extended to the strictly offline imitation setting, and displays qualities of both conservation and adaptability.
We assess AOC's performance in both simulated and real-world healthcare scenarios, emphasizing its capability to manage offline control tasks with high levels of performance while maintaining accountability.
arXiv Detail & Related papers (2023-10-11T17:20:32Z) - Who Audits the Auditors? Recommendations from a field scan of the
algorithmic auditing ecosystem [0.971392598996499]
We provide the first comprehensive field scan of the AI audit ecosystem.
We identify emerging best practices as well as methods and tools that are becoming commonplace.
We outline policy recommendations to improve the quality and impact of these audits.
arXiv Detail & Related papers (2023-10-04T01:40:03Z) - Outsider Oversight: Designing a Third Party Audit Ecosystem for AI
Governance [3.8997087223115634]
We discuss the challenges of third party oversight in the current AI landscape.
We show that the institutional design of such audits are far from monolithic.
We conclude that the turn toward audits alone is unlikely to achieve actual algorithmic accountability.
arXiv Detail & Related papers (2022-06-09T19:18:47Z) - System Cards for AI-Based Decision-Making for Public Policy [5.076419064097733]
This work proposes a system accountability benchmark for formal audits of artificial intelligence-based decision-aiding systems.
It consists of 56 criteria organized within a four-by-four matrix composed of rows focused on (i) data, (ii) model, (iii) code, (iv) system, and columns focused on (a) development, (b) assessment, (c) mitigation, and (d) assurance.
arXiv Detail & Related papers (2022-03-01T18:56:45Z) - Generalizing Cross-Document Event Coreference Resolution Across Multiple
Corpora [63.429307282665704]
Cross-document event coreference resolution (CDCR) is an NLP task in which mentions of events need to be identified and clustered throughout a collection of documents.
CDCR aims to benefit downstream multi-document applications, but improvements from applying CDCR have not been shown yet.
We make the observation that every CDCR system to date was developed, trained, and tested only on a single respective corpus.
arXiv Detail & Related papers (2020-11-24T17:45:03Z) - Towards an Interface Description Template for AI-enabled Systems [77.34726150561087]
Reuse is a common system architecture approach that seeks to instantiate a system architecture with existing components.
There is currently no framework that guides the selection of necessary information to assess their portability to operate in a system different than the one for which the component was originally purposed.
We present ongoing work on establishing an interface description template that captures the main information of an AI-enabled component.
arXiv Detail & Related papers (2020-07-13T20:30:26Z) - Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable
Claims [59.64274607533249]
AI developers need to make verifiable claims to which they can be held accountable.
This report suggests various steps that different stakeholders can take to improve the verifiability of claims made about AI systems.
We analyze ten mechanisms for this purpose--spanning institutions, software, and hardware--and make recommendations aimed at implementing, exploring, or improving those mechanisms.
arXiv Detail & Related papers (2020-04-15T17:15:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.