Householder Dice: A Matrix-Free Algorithm for Simulating Dynamics on
Gaussian and Random Orthogonal Ensembles
- URL: http://arxiv.org/abs/2101.07464v2
- Date: Fri, 22 Jan 2021 00:15:14 GMT
- Title: Householder Dice: A Matrix-Free Algorithm for Simulating Dynamics on
Gaussian and Random Orthogonal Ensembles
- Authors: Yue M. Lu
- Abstract summary: Householder Dice (HD) is an algorithm for simulating dynamics on dense random matrix ensembles with translation-invariant properties.
The memory and costs of the HD algorithm are $mathcalO(nT)$ and $mathcalO(nT2)$, respectively.
Numerical results demonstrate the promise of the HD algorithm as a new computational tool in the study of high-dimensional random systems.
- Score: 12.005731086591139
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper proposes a new algorithm, named Householder Dice (HD), for
simulating dynamics on dense random matrix ensembles with translation-invariant
properties. Examples include the Gaussian ensemble, the Haar-distributed random
orthogonal ensemble, and their complex-valued counterparts. A "direct" approach
to the simulation, where one first generates a dense $n \times n$ matrix from
the ensemble, requires at least $\mathcal{O}(n^2)$ resource in space and time.
The HD algorithm overcomes this $\mathcal{O}(n^2)$ bottleneck by using the
principle of deferred decisions: rather than fixing the entire random matrix in
advance, it lets the randomness unfold with the dynamics. At the heart of this
matrix-free algorithm is an adaptive and recursive construction of (random)
Householder reflectors. These orthogonal transformations exploit the group
symmetry of the matrix ensembles, while simultaneously maintaining the
statistical correlations induced by the dynamics. The memory and computation
costs of the HD algorithm are $\mathcal{O}(nT)$ and $\mathcal{O}(nT^2)$,
respectively, with $T$ being the number of iterations. When $T \ll n$, which is
nearly always the case in practice, the new algorithm leads to significant
reductions in runtime and memory footprint. Numerical results demonstrate the
promise of the HD algorithm as a new computational tool in the study of
high-dimensional random systems.
Related papers
- Fine-grained Analysis and Faster Algorithms for Iteratively Solving Linear Systems [9.30306458153248]
We consider the spectral tail condition number, $kappa_ell$, defined as the ratio between the $ell$th largest and the smallest singular value of the matrix representing the system.
Some of the implications of our result, and of the use of $kappa_ell$, include direct improvement over a fine-grained analysis of the Conjugate method.
arXiv Detail & Related papers (2024-05-09T14:56:49Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Private estimation algorithms for stochastic block models and mixture
models [63.07482515700984]
General tools for designing efficient private estimation algorithms.
First efficient $(epsilon, delta)$-differentially private algorithm for both weak recovery and exact recovery.
arXiv Detail & Related papers (2023-01-11T09:12:28Z) - Multi-block-Single-probe Variance Reduced Estimator for Coupled
Compositional Optimization [49.58290066287418]
We propose a novel method named Multi-block-probe Variance Reduced (MSVR) to alleviate the complexity of compositional problems.
Our results improve upon prior ones in several aspects, including the order of sample complexities and dependence on strongity.
arXiv Detail & Related papers (2022-07-18T12:03:26Z) - Scalable First-Order Bayesian Optimization via Structured Automatic
Differentiation [4.061135251278187]
We show that a wide range of kernels gives rise to structured matrices, enabling an exact $mathcalO(n2d)$ matrix-vector multiply for gradient observations and $mathcalO(n2d2)$ for Hessian observations.
Our methods apply to virtually all canonical kernels and automatically extend to complex kernels, like the neural network, radial basis function network, and spectral mixture kernels.
arXiv Detail & Related papers (2022-06-16T17:59:48Z) - Kernel Packet: An Exact and Scalable Algorithm for Gaussian Process
Regression with Mat\'ern Correlations [23.560067934682294]
We develop an exact and scalable algorithm for one-dimensional Gaussian process regression with Mat'ern correlations.
The proposed algorithm is significantly superior to the existing alternatives in both the computational time and predictive accuracy.
arXiv Detail & Related papers (2022-03-07T03:30:35Z) - Sublinear Time Approximation of Text Similarity Matrices [50.73398637380375]
We introduce a generalization of the popular Nystr"om method to the indefinite setting.
Our algorithm can be applied to any similarity matrix and runs in sublinear time in the size of the matrix.
We show that our method, along with a simple variant of CUR decomposition, performs very well in approximating a variety of similarity matrices.
arXiv Detail & Related papers (2021-12-17T17:04:34Z) - Multiplicative updates for symmetric-cone factorizations [0.0]
We introduce and analyze the symmetric-cone multiplicative update (SCMU) algorithm for computing cone factorizations.
We show that the squared loss objective is non-decreasing along the trajectories of the SCMU algorithm.
arXiv Detail & Related papers (2021-08-02T09:23:39Z) - Quantum algorithms for spectral sums [50.045011844765185]
We propose new quantum algorithms for estimating spectral sums of positive semi-definite (PSD) matrices.
We show how the algorithms and techniques used in this work can be applied to three problems in spectral graph theory.
arXiv Detail & Related papers (2020-11-12T16:29:45Z) - Optimal Iterative Sketching with the Subsampled Randomized Hadamard
Transform [64.90148466525754]
We study the performance of iterative sketching for least-squares problems.
We show that the convergence rate for Haar and randomized Hadamard matrices are identical, andally improve upon random projections.
These techniques may be applied to other algorithms that employ randomized dimension reduction.
arXiv Detail & Related papers (2020-02-03T16:17:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.