Privacy-Enhancing Collaborative Information Sharing through Federated
Learning -- A Case of the Insurance Industry
- URL: http://arxiv.org/abs/2402.14983v1
- Date: Thu, 22 Feb 2024 21:46:24 GMT
- Title: Privacy-Enhancing Collaborative Information Sharing through Federated
Learning -- A Case of the Insurance Industry
- Authors: Panyi Dong, Zhiyu Quan, Brandon Edwards, Shih-han Wang, Runhuan Feng,
Tianyang Wang, Patrick Foley, Prashant Shah
- Abstract summary: The report demonstrates the benefits of harnessing the value of Federated Learning (FL) to learn a single model across multiple insurance industry datasets.
FL addresses two of the most pressing concerns: limited data volume and data variety, which are caused by privacy concerns.
During each round of FL, collaborators compute improvements on the model using their local private data, and these insights are combined to update a global model.
- Score: 1.8092553911119764
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The report demonstrates the benefits (in terms of improved claims loss
modeling) of harnessing the value of Federated Learning (FL) to learn a single
model across multiple insurance industry datasets without requiring the
datasets themselves to be shared from one company to another. The application
of FL addresses two of the most pressing concerns: limited data volume and data
variety, which are caused by privacy concerns, the rarity of claim events, the
lack of informative rating factors, etc.. During each round of FL,
collaborators compute improvements on the model using their local private data,
and these insights are combined to update a global model. Such aggregation of
insights allows for an increase to the effectiveness in forecasting claims
losses compared to models individually trained at each collaborator.
Critically, this approach enables machine learning collaboration without the
need for raw data to leave the compute infrastructure of each respective data
owner. Additionally, the open-source framework, OpenFL, that is used in our
experiments is designed so that it can be run using confidential computing as
well as with additional algorithmic protections against leakage of information
via the shared model updates. In such a way, FL is implemented as a
privacy-enhancing collaborative learning technique that addresses the
challenges posed by the sensitivity and privacy of data in traditional machine
learning solutions. This paper's application of FL can also be expanded to
other areas including fraud detection, catastrophe modeling, etc., that have a
similar need to incorporate data privacy into machine learning collaborations.
Our framework and empirical results provide a foundation for future
collaborations among insurers, regulators, academic researchers, and InsurTech
experts.
Related papers
- Privacy in Federated Learning [0.0]
Federated Learning (FL) represents a significant advancement in distributed machine learning.
This chapter delves into the core privacy concerns within FL, including the risks of data reconstruction, model inversion attacks, and membership inference.
It examines the trade-offs between model accuracy and privacy, emphasizing the importance of balancing these factors in practical implementations.
arXiv Detail & Related papers (2024-08-12T18:41:58Z) - Federated Knowledge Graph Unlearning via Diffusion Model [5.373752180709173]
Federated learning (FL) promotes the development and application of artificial intelligence technologies.
In this paper, we propose FedDM, a novel framework tailored for machine unlearning in federated knowledge graphs.
arXiv Detail & Related papers (2024-03-13T14:06:51Z) - FewFedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning [54.26614091429253]
Federated instruction tuning (FedIT) is a promising solution, by consolidating collaborative training across multiple data owners.
FedIT encounters limitations such as scarcity of instructional data and risk of exposure to training data extraction attacks.
We propose FewFedPIT, designed to simultaneously enhance privacy protection and model performance of federated few-shot learning.
arXiv Detail & Related papers (2024-03-10T08:41:22Z) - Personalized Federated Learning with Attention-based Client Selection [57.71009302168411]
We propose FedACS, a new PFL algorithm with an Attention-based Client Selection mechanism.
FedACS integrates an attention mechanism to enhance collaboration among clients with similar data distributions.
Experiments on CIFAR10 and FMNIST validate FedACS's superiority.
arXiv Detail & Related papers (2023-12-23T03:31:46Z) - Federated Learning Empowered by Generative Content [55.576885852501775]
Federated learning (FL) enables leveraging distributed private data for model training in a privacy-preserving way.
We propose a novel FL framework termed FedGC, designed to mitigate data heterogeneity issues by diversifying private data with generative content.
We conduct a systematic empirical study on FedGC, covering diverse baselines, datasets, scenarios, and modalities.
arXiv Detail & Related papers (2023-12-10T07:38:56Z) - Privacy-preserving Federated Primal-dual Learning for Non-convex and Non-smooth Problems with Model Sparsification [51.04894019092156]
Federated learning (FL) has been recognized as a rapidly growing area, where the model is trained over clients under the FL orchestration (PS)
In this paper, we propose a novel primal sparification algorithm for and guarantee non-smooth FL problems.
Its unique insightful properties and its analyses are also presented.
arXiv Detail & Related papers (2023-10-30T14:15:47Z) - Federated Learning with Reduced Information Leakage and Computation [17.069452700698047]
Federated learning (FL) is a distributed learning paradigm that allows multiple decentralized clients to collaboratively learn a common model without sharing local data.
This paper introduces Upcycled-FL, a strategy that applies first-order approximation at every even round of model update.
Under this strategy, half of the FL updates incur no information leakage and require much less computational and transmission costs.
arXiv Detail & Related papers (2023-10-10T06:22:06Z) - PS-FedGAN: An Efficient Federated Learning Framework Based on Partially
Shared Generative Adversarial Networks For Data Privacy [56.347786940414935]
Federated Learning (FL) has emerged as an effective learning paradigm for distributed computation.
This work proposes a novel FL framework that requires only partial GAN model sharing.
Named as PS-FedGAN, this new framework enhances the GAN releasing and training mechanism to address heterogeneous data distributions.
arXiv Detail & Related papers (2023-05-19T05:39:40Z) - Benchmarking FedAvg and FedCurv for Image Classification Tasks [1.376408511310322]
This paper focuses on the problem of statistical heterogeneity of the data in the same federated network.
Several Federated Learning algorithms, such as FedAvg, FedProx and Federated Curvature (FedCurv) have already been proposed.
As a side product of this work, we release the non-IID version of the datasets we used so to facilitate further comparisons from the FL community.
arXiv Detail & Related papers (2023-03-31T10:13:01Z) - Distributed Machine Learning and the Semblance of Trust [66.1227776348216]
Federated Learning (FL) allows the data owner to maintain data governance and perform model training locally without having to share their data.
FL and related techniques are often described as privacy-preserving.
We explain why this term is not appropriate and outline the risks associated with over-reliance on protocols that were not designed with formal definitions of privacy in mind.
arXiv Detail & Related papers (2021-12-21T08:44:05Z) - Privacy Preservation in Federated Learning: An insightful survey from
the GDPR Perspective [10.901568085406753]
Article is dedicated to surveying on the state-of-the-art privacy techniques, which can be employed in Federated learning.
Recent research has demonstrated that retaining data and on computation in FL is not enough for privacy-guarantee.
This is because ML model parameters exchanged between parties in an FL system, which can be exploited in some privacy attacks.
arXiv Detail & Related papers (2020-11-10T21:41:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.