Federated Learning Priorities Under the European Union Artificial
Intelligence Act
- URL: http://arxiv.org/abs/2402.05968v1
- Date: Mon, 5 Feb 2024 19:52:19 GMT
- Title: Federated Learning Priorities Under the European Union Artificial
Intelligence Act
- Authors: Herbert Woisetschl\"ager, Alexander Erben, Bill Marino, Shiqiang Wang,
Nicholas D. Lane, Ruben Mayer, Hans-Arno Jacobsen
- Abstract summary: We perform a first-of-its-kind interdisciplinary analysis (legal and ML) of the impact the AI Act may have on Federated Learning.
We explore data governance issues and the concern for privacy.
Most noteworthy are the opportunities to defend against data bias and enhance private and secure computation.
- Score: 68.44894319552114
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The age of AI regulation is upon us, with the European Union Artificial
Intelligence Act (AI Act) leading the way. Our key inquiry is how this will
affect Federated Learning (FL), whose starting point of prioritizing data
privacy while performing ML fundamentally differs from that of centralized
learning. We believe the AI Act and future regulations could be the missing
catalyst that pushes FL toward mainstream adoption. However, this can only
occur if the FL community reprioritizes its research focus. In our position
paper, we perform a first-of-its-kind interdisciplinary analysis (legal and ML)
of the impact the AI Act may have on FL and make a series of observations
supporting our primary position through quantitative and qualitative analysis.
We explore data governance issues and the concern for privacy. We establish new
challenges regarding performance and energy efficiency within lifecycle
monitoring. Taken together, our analysis suggests there is a sizable
opportunity for FL to become a crucial component of AI Act-compliant ML systems
and for the new regulation to drive the adoption of FL techniques in general.
Most noteworthy are the opportunities to defend against data bias and enhance
private and secure computation
Related papers
- Using AI Alignment Theory to understand the potential pitfalls of regulatory frameworks [55.2480439325792]
This paper critically examines the European Union's Artificial Intelligence Act (EU AI Act)
Uses insights from Alignment Theory (AT) research, which focuses on the potential pitfalls of technical alignment in Artificial Intelligence.
As we apply these concepts to the EU AI Act, we uncover potential vulnerabilities and areas for improvement in the regulation.
arXiv Detail & Related papers (2024-10-10T17:38:38Z) - Federated Learning and AI Regulation in the European Union: Who is Responsible? -- An Interdisciplinary Analysis [12.098213443468346]
The European Union Artificial Intelligence Act mandates clear stakeholder responsibilities in developing and deploying machine learning applications.
Federated Learning enables the training of generative AI Models across data siloes, sharing only model parameters while improving data security.
Our work contributes to clarifying the roles of both parties, explains strategies for shifting responsibilities to the server operator, and points out open technical challenges.
arXiv Detail & Related papers (2024-07-11T00:41:16Z) - The risks of risk-based AI regulation: taking liability seriously [46.90451304069951]
The development and regulation of AI seems to have reached a critical stage.
Some experts are calling for a moratorium on the training of AI systems more powerful than GPT-4.
This paper analyses the most advanced legal proposal, the European Union's AI Act.
arXiv Detail & Related papers (2023-11-03T12:51:37Z) - A Survey of Federated Unlearning: A Taxonomy, Challenges and Future
Directions [71.16718184611673]
The evolution of privacy-preserving Federated Learning (FL) has led to an increasing demand for implementing the right to be forgotten.
The implementation of selective forgetting is particularly challenging in FL due to its decentralized nature.
Federated Unlearning (FU) emerges as a strategic solution to address the increasing need for data privacy.
arXiv Detail & Related papers (2023-10-30T01:34:33Z) - Towards Interpretable Federated Learning [19.764172768506132]
Federated learning (FL) enables multiple data owners to build machine learning models collaboratively without exposing their private local data.
It is important to balance the need for performance, privacy-preservation and interpretability, especially in mission critical applications such as finance and healthcare.
We conduct comprehensive analysis of the representative IFL approaches, the commonly adopted performance evaluation metrics, and promising directions towards building versatile IFL techniques.
arXiv Detail & Related papers (2023-02-27T02:06:18Z) - Towards an Accountable and Reproducible Federated Learning: A FactSheets
Approach [6.488712018186561]
Federated Learning (FL) is a novel paradigm for the shared training of models based on decentralized and private data.
We introduce AF2 Framework, where we instrument FL with accountability by fusing verifiable claims with tamper-evident facts.
We build on AI FactSheets for instilling transparency and trustworthiness into the AI lifecycle.
arXiv Detail & Related papers (2022-02-25T00:34:14Z) - Privacy and Robustness in Federated Learning: Attacks and Defenses [74.62641494122988]
We conduct the first comprehensive survey on this topic.
Through a concise introduction to the concept of FL, and a unique taxonomy covering: 1) threat models; 2) poisoning attacks and defenses against robustness; 3) inference attacks and defenses against privacy, we provide an accessible review of this important topic.
arXiv Detail & Related papers (2020-12-07T12:11:45Z) - Trustworthy AI [75.99046162669997]
Brittleness to minor adversarial changes in the input data, ability to explain the decisions, address the bias in their training data, are some of the most prominent limitations.
We propose the tutorial on Trustworthy AI to address six critical issues in enhancing user and public trust in AI systems.
arXiv Detail & Related papers (2020-11-02T20:04:18Z) - Bias in Data-driven AI Systems -- An Introductory Survey [37.34717604783343]
This survey focuses on data-driven AI, as a large part of AI is powered nowadays by (big) data and powerful Machine Learning (ML) algorithms.
If otherwise not specified, we use the general term bias to describe problems related to the gathering or processing of data that might result in prejudiced decisions on the bases of demographic features like race, sex, etc.
arXiv Detail & Related papers (2020-01-14T09:39:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.