Communication-Efficient Federated Linear and Deep Generalized Canonical
Correlation Analysis
- URL: http://arxiv.org/abs/2109.12400v2
- Date: Mon, 3 Apr 2023 19:17:47 GMT
- Title: Communication-Efficient Federated Linear and Deep Generalized Canonical
Correlation Analysis
- Authors: Sagar Shrestha and Xiao Fu
- Abstract summary: This work puts forth a communication-efficient federated learning framework for both linear and deep GCCA.
Compared to the unquantized version, our empirical study shows that the proposed algorithm enjoys a substantial reduction of communication overheads with virtually no loss in accuracy and convergence speed.
- Score: 13.04301271535511
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classic and deep generalized canonical correlation analysis (GCCA) algorithms
seek low-dimensional common representations of data entities from multiple
``views'' (e.g., audio and image) using linear transformations and neural
networks, respectively. When the views are acquired and stored at different
computing agents (e.g., organizations and edge devices) and data sharing is
undesired due to privacy or communication cost considerations, federated
learning-based GCCA is well-motivated. In federated learning, the views are
kept locally at the agents and only derived, limited information exchange with
a central server is allowed. However, applying existing GCCA algorithms onto
such federated learning settings may incur prohibitively high communication
overhead. This work puts forth a communication-efficient federated learning
framework for both linear and deep GCCA under the maximum variance (MAX-VAR)
formulation. The overhead issue is addressed by aggressively compressing (via
quantization) the exchanging information between the computing agents and a
central controller. Compared to the unquantized version, our empirical study
shows that the proposed algorithm enjoys a substantial reduction of
communication overheads with virtually no loss in accuracy and convergence
speed. Rigorous convergence analyses are also presented, which is a nontrivial
effort. Generic federated optimization results do not cover the special problem
structure of GCCA. Our result shows that the proposed algorithms for both
linear and deep GCCA converge to critical points at a sublinear rate, even
under heavy quantization and stochastic approximations. In addition, in the
linear MAX-VAR case, the quantized algorithm approaches a global optimum in a
geometric rate under reasonable conditions. Synthetic and real-data experiments
are used to showcase the effectiveness of the proposed approach.
Related papers
- Boosting the Performance of Decentralized Federated Learning via Catalyst Acceleration [66.43954501171292]
We introduce Catalyst Acceleration and propose an acceleration Decentralized Federated Learning algorithm called DFedCata.
DFedCata consists of two main components: the Moreau envelope function, which addresses parameter inconsistencies, and Nesterov's extrapolation step, which accelerates the aggregation phase.
Empirically, we demonstrate the advantages of the proposed algorithm in both convergence speed and generalization performance on CIFAR10/100 with various non-iid data distributions.
arXiv Detail & Related papers (2024-10-09T06:17:16Z) - Communication-Efficient Federated Non-Linear Bandit Optimization [26.23638987873429]
We propose a new algorithm, named Fed-GO-UCB, for federated bandit optimization with generic non-linear objective function.
Under some mild conditions, we rigorously prove that Fed-GO-UCB is able to achieve sub-linear rate for both cumulative regret and communication cost.
arXiv Detail & Related papers (2023-11-03T03:50:31Z) - Domain Adaptation Principal Component Analysis: base linear method for
learning with out-of-distribution data [55.41644538483948]
Domain adaptation is a popular paradigm in modern machine learning.
We present a method called Domain Adaptation Principal Component Analysis (DAPCA)
DAPCA finds a linear reduced data representation useful for solving the domain adaptation task.
arXiv Detail & Related papers (2022-08-28T21:10:56Z) - Task-Oriented Sensing, Computation, and Communication Integration for
Multi-Device Edge AI [108.08079323459822]
This paper studies a new multi-intelligent edge artificial-latency (AI) system, which jointly exploits the AI model split inference and integrated sensing and communication (ISAC)
We measure the inference accuracy by adopting an approximate but tractable metric, namely discriminant gain.
arXiv Detail & Related papers (2022-07-03T06:57:07Z) - On Leave-One-Out Conditional Mutual Information For Generalization [122.2734338600665]
We derive information theoretic generalization bounds for supervised learning algorithms based on a new measure of leave-one-out conditional mutual information (loo-CMI)
Contrary to other CMI bounds, our loo-CMI bounds can be computed easily and can be interpreted in connection to other notions such as classical leave-one-out cross-validation.
We empirically validate the quality of the bound by evaluating its predicted generalization gap in scenarios for deep learning.
arXiv Detail & Related papers (2022-07-01T17:58:29Z) - A Newton-type algorithm for federated learning based on incremental
Hessian eigenvector sharing [5.404315085380945]
We present an original communication-constrained Newton-type (NT) algorithm designed to accelerate Federated Learning (FL)
The proposed solution is thoroughly validated on real datasets.
arXiv Detail & Related papers (2022-02-11T17:52:56Z) - Resource-constrained Federated Edge Learning with Heterogeneous Data:
Formulation and Analysis [8.863089484787835]
We propose a distributed approximate Newton-type Newton-type training scheme, namely FedOVA, to solve the heterogeneous statistical challenge brought by heterogeneous data.
FedOVA decomposes a multi-class classification problem into more straightforward binary classification problems and then combines their respective outputs using ensemble learning.
arXiv Detail & Related papers (2021-10-14T17:35:24Z) - Sample-based and Feature-based Federated Learning via Mini-batch SSCA [18.11773963976481]
This paper investigates sample-based and feature-based federated optimization.
We show that the proposed algorithms can preserve data privacy through the model aggregation mechanism.
We also show that the proposed algorithms converge to Karush-Kuhn-Tucker points of the respective federated optimization problems.
arXiv Detail & Related papers (2021-04-13T08:23:46Z) - FedPD: A Federated Learning Framework with Optimal Rates and Adaptivity
to Non-IID Data [59.50904660420082]
Federated Learning (FL) has become a popular paradigm for learning from distributed data.
To effectively utilize data at different devices without moving them to the cloud, algorithms such as the Federated Averaging (FedAvg) have adopted a "computation then aggregation" (CTA) model.
arXiv Detail & Related papers (2020-05-22T23:07:42Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - Generalized Canonical Correlation Analysis: A Subspace Intersection
Approach [30.475159163815505]
Generalized Canonical Correlation Analysis (GCCA) is an important tool that finds numerous applications in data mining, machine learning, and artificial intelligence.
This paper offers a fresh algebraic perspective of GCCA based on a (bi-linear) generative model that naturally captures its essence.
A novel GCCA algorithm is proposed based on subspace intersection, which scales up to handle large GCCA tasks.
arXiv Detail & Related papers (2020-03-25T04:04:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.