Leveraging Spatial and Temporal Correlations in Sparsified Mean
Estimation
- URL: http://arxiv.org/abs/2110.07751v1
- Date: Thu, 14 Oct 2021 22:24:26 GMT
- Title: Leveraging Spatial and Temporal Correlations in Sparsified Mean
Estimation
- Authors: Divyansh Jhunjhunwala, Ankur Mallick, Advait Gadhikar, Swanand Kadhe,
Gauri Joshi
- Abstract summary: We study the problem of estimating at a central server the mean of a set of vectors distributed across several nodes.
We leverage these correlations by simply modifying the decoding method used by the server to estimate the mean.
We provide an analysis of the resulting estimation error as well as experiments for PCA, K-Means and Logistic Regression.
- Score: 11.602121447683597
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of estimating at a central server the mean of a set of
vectors distributed across several nodes (one vector per node). When the
vectors are high-dimensional, the communication cost of sending entire vectors
may be prohibitive, and it may be imperative for them to use sparsification
techniques. While most existing work on sparsified mean estimation is agnostic
to the characteristics of the data vectors, in many practical applications such
as federated learning, there may be spatial correlations (similarities in the
vectors sent by different nodes) or temporal correlations (similarities in the
data sent by a single node over different iterations of the algorithm) in the
data vectors. We leverage these correlations by simply modifying the decoding
method used by the server to estimate the mean. We provide an analysis of the
resulting estimation error as well as experiments for PCA, K-Means and Logistic
Regression, which show that our estimators consistently outperform more
sophisticated and expensive sparsification methods.
Related papers
- Semiparametric conformal prediction [79.6147286161434]
Risk-sensitive applications require well-calibrated prediction sets over multiple, potentially correlated target variables.
We treat the scores as random vectors and aim to construct the prediction set accounting for their joint correlation structure.
We report desired coverage and competitive efficiency on a range of real-world regression problems.
arXiv Detail & Related papers (2024-11-04T14:29:02Z) - Improving Out-of-Distribution Generalization of Neural Rerankers with
Contextualized Late Interaction [52.63663547523033]
Late interaction, the simplest form of multi-vector, is also helpful to neural rerankers that only use the [] vector to compute the similarity score.
We show that the finding is consistent across different model sizes and first-stage retrievers of diverse natures.
arXiv Detail & Related papers (2023-02-13T18:42:17Z) - Multi-Fidelity Covariance Estimation in the Log-Euclidean Geometry [0.0]
We introduce a multi-fidelity estimator of covariance matrices that employs the log-Euclidean geometry of the symmetric positive-definite manifold.
We develop an optimal sample allocation scheme that minimizes the mean-squared error of the estimator given a fixed budget.
Evaluations of our approach using data from physical applications demonstrate more accurate metric learning and speedups of more than one order of magnitude compared to benchmarks.
arXiv Detail & Related papers (2023-01-31T16:33:46Z) - Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic
Uncertainty [58.144520501201995]
Bi-Lipschitz regularization of neural network layers preserve relative distances between data instances in the feature spaces of each layer.
With the use of an attentive set encoder, we propose to meta learn either diagonal or diagonal plus low-rank factors to efficiently construct task specific covariance matrices.
We also propose an inference procedure which utilizes scaled energy to achieve a final predictive distribution.
arXiv Detail & Related papers (2021-10-12T22:04:19Z) - Robust Regularized Locality Preserving Indexing for Fiedler Vector
Estimation [32.26669925809068]
In real-world applications, the data may be subject to heavy-tailed noise and outliers which results in deteriorations in the structure of the Fiedler vector estimate.
We design a Robust Regularized Locality Preserving Indexing (RRLPI) method for Fiedler vector estimation that aims to approximate the nonlinear manifold structure of the Laplace Beltrami operator.
arXiv Detail & Related papers (2021-07-26T09:49:23Z) - Learning Optical Flow from a Few Matches [67.83633948984954]
We show that the dense correlation volume representation is redundant and accurate flow estimation can be achieved with only a fraction of elements in it.
Experiments show that our method can reduce computational cost and memory use significantly, while maintaining high accuracy.
arXiv Detail & Related papers (2021-04-05T21:44:00Z) - Scalable Vector Gaussian Information Bottleneck [19.21005180893519]
We study a variation of the problem, called scalable information bottleneck, in which the encoder outputs multiple descriptions of the observation.
We derive a variational inference type algorithm for general sources with unknown distribution; and show means of parametrizing it using neural networks.
arXiv Detail & Related papers (2021-02-15T12:51:26Z) - Wyner-Ziv Estimators: Efficient Distributed Mean Estimation with Side
Information [36.49502352300087]
We study the problem of distributed mean estimation where the server has access to side information.
We propose emphWyner-Ziv estimators, which are communication and computationally efficient.
In a different direction, we present an alternative Wyner-Ziv estimator that uses correlated sampling.
arXiv Detail & Related papers (2020-11-24T15:19:55Z) - Evaluating representations by the complexity of learning low-loss
predictors [55.94170724668857]
We consider the problem of evaluating representations of data for use in solving a downstream task.
We propose to measure the quality of a representation by the complexity of learning a predictor on top of the representation that achieves low loss on a task of interest.
arXiv Detail & Related papers (2020-09-15T22:06:58Z) - $\gamma$-ABC: Outlier-Robust Approximate Bayesian Computation Based on a
Robust Divergence Estimator [95.71091446753414]
We propose to use a nearest-neighbor-based $gamma$-divergence estimator as a data discrepancy measure.
Our method achieves significantly higher robustness than existing discrepancy measures.
arXiv Detail & Related papers (2020-06-13T06:09:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.