Knowledge-Injected Federated Learning
- URL: http://arxiv.org/abs/2208.07530v1
- Date: Tue, 16 Aug 2022 04:23:06 GMT
- Title: Knowledge-Injected Federated Learning
- Authors: Zhenan Fan, Zirui Zhou, Jian Pei, Michael P. Friedlander, Jiajie Hu,
Chengliang Li, Yong Zhang
- Abstract summary: Federated learning is an emerging technique for training models from decentralized data sets.
In many applications, data owners participating in the federated learning system hold not only the data but also a set of domain knowledge.
We propose a federated learning framework that allows the injection of participants' domain knowledge.
- Score: 44.89926234630289
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning is an emerging technique for training models from
decentralized data sets. In many applications, data owners participating in the
federated learning system hold not only the data but also a set of domain
knowledge. Such knowledge includes human know-how and craftsmanship that can be
extremely helpful to the federated learning task. In this work, we propose a
federated learning framework that allows the injection of participants' domain
knowledge, where the key idea is to refine the global model with knowledge
locally. The scenario we consider is motivated by a real industry-level
application, and we demonstrate the effectiveness of our approach to this
application.
Related papers
- LEKA:LLM-Enhanced Knowledge Augmentation [24.552995956148145]
Humans excel in analogical learning and knowledge transfer.
Models would transition from passively acquiring to actively accessing and learning from knowledge.
We develop a knowledge augmentation method LEKA for knowledge transfer.
arXiv Detail & Related papers (2025-01-29T17:44:57Z) - Large Language Models are Limited in Out-of-Context Knowledge Reasoning [65.72847298578071]
Large Language Models (LLMs) possess extensive knowledge and strong capabilities in performing in-context reasoning.
This paper focuses on a significant aspect of out-of-context reasoning: Out-of-Context Knowledge Reasoning (OCKR), which is to combine multiple knowledge to infer new knowledge.
arXiv Detail & Related papers (2024-06-11T15:58:59Z) - KnFu: Effective Knowledge Fusion [5.305607095162403]
Federated Learning (FL) has emerged as a prominent alternative to the traditional centralized learning approach.
The paper proposes Effective Knowledge Fusion (KnFu) algorithm that evaluates knowledge of local models to only fuse semantic neighbors' effective knowledge for each client.
A key conclusion of the work is that in scenarios with large and highly heterogeneous local datasets, local training could be preferable to knowledge fusion-based solutions.
arXiv Detail & Related papers (2024-03-18T15:49:48Z) - Private Knowledge Sharing in Distributed Learning: A Survey [50.51431815732716]
The rise of Artificial Intelligence has revolutionized numerous industries and transformed the way society operates.
It is crucial to utilize information in learning processes that are either distributed or owned by different entities.
Modern data-driven services have been developed to integrate distributed knowledge entities into their outcomes.
arXiv Detail & Related papers (2024-02-08T07:18:23Z) - Exploring Federated Unlearning: Analysis, Comparison, and Insights [101.64910079905566]
federated unlearning enables the selective removal of data from models trained in federated systems.
This paper examines existing federated unlearning approaches, examining their algorithmic efficiency, impact on model accuracy, and effectiveness in preserving privacy.
We propose the OpenFederatedUnlearning framework, a unified benchmark for evaluating federated unlearning methods.
arXiv Detail & Related papers (2023-10-30T01:34:33Z) - Federated Learning: Organizational Opportunities, Challenges, and
Adoption Strategies [39.58317527488534]
Federated learning allows distributed clients to train models collaboratively without the need to share their respective training data with others.
We argue that federated learning presents organizational challenges with ample interdisciplinary opportunities for information systems researchers.
arXiv Detail & Related papers (2023-08-04T09:23:23Z) - Heterogeneous Federated Knowledge Graph Embedding Learning and
Unlearning [14.063276595895049]
Federated Learning (FL) is a paradigm to train a global machine learning model across distributed clients without sharing raw data.
We propose FedLU, a novel FL framework for heterogeneous KG embedding learning and unlearning.
We show that FedLU achieves superior results in both link prediction and knowledge forgetting.
arXiv Detail & Related papers (2023-02-04T02:44:48Z) - Knowledge-augmented Deep Learning and Its Applications: A Survey [60.221292040710885]
knowledge-augmented deep learning (KADL) aims to identify domain knowledge and integrate it into deep models for data-efficient, generalizable, and interpretable deep learning.
This survey subsumes existing works and offers a bird's-eye view of research in the general area of knowledge-augmented deep learning.
arXiv Detail & Related papers (2022-11-30T03:44:15Z) - Knowledge-Aware Federated Active Learning with Non-IID Data [75.98707107158175]
We propose a federated active learning paradigm to efficiently learn a global model with limited annotation budget.
The main challenge faced by federated active learning is the mismatch between the active sampling goal of the global model on the server and that of the local clients.
We propose Knowledge-Aware Federated Active Learning (KAFAL), which consists of Knowledge-Specialized Active Sampling (KSAS) and Knowledge-Compensatory Federated Update (KCFU)
arXiv Detail & Related papers (2022-11-24T13:08:43Z) - Global Knowledge Distillation in Federated Learning [3.7311680121118345]
We propose a novel global knowledge distillation method, named FedGKD, which learns the knowledge from past global models to tackle down the local bias training problem.
To demonstrate the effectiveness of the proposed method, we conduct extensive experiments on various CV datasets (CIFAR-10/100) and settings (non-i.i.d data)
arXiv Detail & Related papers (2021-06-30T18:14:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.