Desiderata for Explainable AI in statistical production systems of the
European Central Bank
- URL: http://arxiv.org/abs/2107.08045v1
- Date: Sun, 18 Jul 2021 05:58:11 GMT
- Title: Desiderata for Explainable AI in statistical production systems of the
European Central Bank
- Authors: Carlos Mougan Navarro, Georgios Kanellos, Thomas Gottron
- Abstract summary: We aim to state user-centric desiderata for explainable AI reflecting common explainability needs experienced in statistical production systems of the European Central Bank.
We provide two concrete use cases from the domain of statistical data production in central banks: the detection of outliers in the Centralised Securities Database and the data-driven identification of data quality checks for the Supervisory Banking data system.
- Score: 0.537133760455631
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Explainable AI constitutes a fundamental step towards establishing fairness
and addressing bias in algorithmic decision-making. Despite the large body of
work on the topic, the benefit of solutions is mostly evaluated from a
conceptual or theoretical point of view and the usefulness for real-world use
cases remains uncertain. In this work, we aim to state clear user-centric
desiderata for explainable AI reflecting common explainability needs
experienced in statistical production systems of the European Central Bank. We
link the desiderata to archetypical user roles and give examples of techniques
and methods which can be used to address the user's needs. To this end, we
provide two concrete use cases from the domain of statistical data production
in central banks: the detection of outliers in the Centralised Securities
Database and the data-driven identification of data quality checks for the
Supervisory Banking data system.
Related papers
- Stream-Based Monitoring of Algorithmic Fairness [4.811789437743092]
Stream-based monitoring is proposed as a solution for verifying the algorithmic fairness of decision and prediction systems at runtime.
We present a principled way to formalize algorithmic fairness over temporal data streams in the specification language RTLola.
arXiv Detail & Related papers (2025-01-30T13:18:59Z) - Demographic Benchmarking: Bridging Socio-Technical Gaps in Bias Detection [0.0]
This paper describes how the ITACA AI auditing platform tackles demographic benchmarking when auditing AI recommender systems.
The framework serves us as auditors as it allows us to not just measure but establish acceptability ranges for specific performance indicators.
Our approach integrates socio-demographic insights directly into AI systems, reducing bias and improving overall performance.
arXiv Detail & Related papers (2025-01-27T12:14:49Z) - Pessimistic Evaluation [58.736490198613154]
We argue that evaluating information access systems assumes utilitarian values not aligned with traditions of information access based on equal access.
We advocate for pessimistic evaluation of information access systems focusing on worst case utility.
arXiv Detail & Related papers (2024-10-17T15:40:09Z) - InsightBench: Evaluating Business Analytics Agents Through Multi-Step Insight Generation [79.09622602860703]
We introduce InsightBench, a benchmark dataset with three key features.
It consists of 100 datasets representing diverse business use cases such as finance and incident management.
Unlike existing benchmarks focusing on answering single queries, InsightBench evaluates agents based on their ability to perform end-to-end data analytics.
arXiv Detail & Related papers (2024-07-08T22:06:09Z) - Human-Centric Multimodal Machine Learning: Recent Advances and Testbed
on AI-based Recruitment [66.91538273487379]
There is a certain consensus about the need to develop AI applications with a Human-Centric approach.
Human-Centric Machine Learning needs to be developed based on four main requirements: (i) utility and social good; (ii) privacy and data ownership; (iii) transparency and accountability; and (iv) fairness in AI-driven decision-making processes.
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
arXiv Detail & Related papers (2023-02-13T16:44:44Z) - NLP-based Decision Support System for Examination of Eligibility
Criteria from Securities Prospectuses at the German Central Bank [0.0]
The German Central Bank receives hundreds of scanned prospectuses in PDF format, which must be manually processed to decide upon their eligibility.
We found that this tedious and time-consuming process can be (semi-automated) by employing modern NLP model architectures.
The proposed Decision Support System provides decisions of document-level eligibility criteria accompanied by human-understandable explanations.
arXiv Detail & Related papers (2023-02-09T11:00:58Z) - Introducing explainable supervised machine learning into interactive
feedback loops for statistical production system [0.13999481573773068]
We develop an interactive feedback loop between data collected by the European Central Bank and data quality assurance performed by National Central Banks.
The feedback loop is based on a set of rule-based checks for raising exceptions, upon which the user either confirms the data or corrects an actual error.
In this paper we use the information received from this feedback loop to optimize the exceptions presented to the National Central Banks.
arXiv Detail & Related papers (2022-02-07T14:17:06Z) - AI Assurance using Causal Inference: Application to Public Policy [0.0]
Most AI approaches can only be represented as "black boxes" and suffer from the lack of transparency.
It is crucial not only to develop effective and robust AI systems, but to make sure their internal processes are explainable and fair.
arXiv Detail & Related papers (2021-12-01T16:03:06Z) - Multisource AI Scorecard Table for System Evaluation [3.74397577716445]
The paper describes a Multisource AI Scorecard Table (MAST) that provides the developer and user of an artificial intelligence (AI)/machine learning (ML) system with a standard checklist.
The paper explores how the analytic tradecraft standards outlined in Intelligence Community Directive (ICD) 203 can provide a framework for assessing the performance of an AI system.
arXiv Detail & Related papers (2021-02-08T03:37:40Z) - Leveraging Expert Consistency to Improve Algorithmic Decision Support [62.61153549123407]
We explore the use of historical expert decisions as a rich source of information that can be combined with observed outcomes to narrow the construct gap.
We propose an influence function-based methodology to estimate expert consistency indirectly when each case in the data is assessed by a single expert.
Our empirical evaluation, using simulations in a clinical setting and real-world data from the child welfare domain, indicates that the proposed approach successfully narrows the construct gap.
arXiv Detail & Related papers (2021-01-24T05:40:29Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z) - Super-App Behavioral Patterns in Credit Risk Models: Financial,
Statistical and Regulatory Implications [110.54266632357673]
We present the impact of alternative data that originates from an app-based marketplace, in contrast to traditional bureau data, upon credit scoring models.
Our results, validated across two countries, show that these new sources of data are particularly useful for predicting financial behavior in low-wealth and young individuals.
arXiv Detail & Related papers (2020-05-09T01:32:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.