Fast optimization of common basis for matrix set through Common Singular
Value Decomposition
- URL: http://arxiv.org/abs/2204.08242v1
- Date: Mon, 18 Apr 2022 10:18:51 GMT
- Title: Fast optimization of common basis for matrix set through Common Singular
Value Decomposition
- Authors: Jarek Duda
- Abstract summary: Proposed CSVD (common SVD): fast general approach based on SVD.
$U$ as built of eigenvectors of $sum_i (w_k)q (A_k A_kT)p$ and $V$ of $sum_k (w_k)q (A_kT A_k)p$, where $w_k$ are their weights.
- Score: 0.8702432681310399
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: SVD (singular value decomposition) is one of the basic tools of machine
learning, allowing to optimize basis for a given matrix. However, sometimes we
have a set of matrices $\{A_k\}_k$ instead, and would like to optimize a single
common basis for them: find orthogonal matrices $U$, $V$, such that $\{U^T A_k
V\}$ set of matrices is somehow simpler. For example DCT-II is orthonormal
basis of functions commonly used in image/video compression - as discussed
here, this kind of basis can be quickly automatically optimized for a given
dataset. While also discussed gradient descent optimization might be
computationally costly, there is proposed CSVD (common SVD): fast general
approach based on SVD. Specifically, we choose $U$ as built of eigenvectors of
$\sum_i (w_k)^q (A_k A_k^T)^p$ and $V$ of $\sum_k (w_k)^q (A_k^T A_k)^p$, where
$w_k$ are their weights, $p,q>0$ are some chosen powers e.g. 1/2, optionally
with normalization e.g. $A \to A - rc^T$ where $r_i=\sum_j A_{ij}, c_j =\sum_i
A_{ij}$.
Related papers
- Query Efficient Structured Matrix Learning [32.0553563150929]
We find a near-optimal approximation to $A$ from any finite-sized family of matrices, $mathcalF$.<n>Surprisingly, we show that, in the matvec model, it is possible to obtain a nearly quadratic improvement in complexity, to $tildeO(sqrtlog|mathcalF|)$.<n>As an example, we establish that a near-optimal approximation from any emphlinear matrix family of dimension $q$ can be learned with $tildeO(sqrt
arXiv Detail & Related papers (2025-07-25T14:04:20Z) - Optimal Quantization for Matrix Multiplication [35.007966885532724]
We present a universal quantizer based on nested lattices with an explicit guarantee of approximation error for any (non-random) pair of matrices $A$, $B$ in terms of only Frobenius norms $|A|_F, |B|_F$ and $|Atop B|_F$.
arXiv Detail & Related papers (2024-10-17T17:19:48Z) - Optimal Sketching for Residual Error Estimation for Matrix and Vector Norms [50.15964512954274]
We study the problem of residual error estimation for matrix and vector norms using a linear sketch.
We demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work.
We also show an $Omega(k2/pn1-2/p)$ lower bound for the sparse recovery problem, which is tight up to a $mathrmpoly(log n)$ factor.
arXiv Detail & Related papers (2024-08-16T02:33:07Z) - Data-Driven Linear Complexity Low-Rank Approximation of General Kernel
Matrices: A Geometric Approach [0.9453554184019107]
A kernel matrix may be defined as $K_ij = kappa(x_i,y_j)$ where $kappa(x,y)$ is a kernel function.
We seek a low-rank approximation to a kernel matrix where the sets of points $X$ and $Y$ are large.
arXiv Detail & Related papers (2022-12-24T07:15:00Z) - Optimal Query Complexities for Dynamic Trace Estimation [59.032228008383484]
We consider the problem of minimizing the number of matrix-vector queries needed for accurate trace estimation in the dynamic setting where our underlying matrix is changing slowly.
We provide a novel binary tree summation procedure that simultaneously estimates all $m$ traces up to $epsilon$ error with $delta$ failure probability.
Our lower bounds (1) give the first tight bounds for Hutchinson's estimator in the matrix-vector product model with Frobenius norm error even in the static setting, and (2) are the first unconditional lower bounds for dynamic trace estimation.
arXiv Detail & Related papers (2022-09-30T04:15:44Z) - Private Matrix Approximation and Geometry of Unitary Orbits [29.072423395363668]
This problem seeks to approximate $A$ by a matrix whose spectrum is the same as $Lambda$.
We give efficient and private algorithms that come with upper and lower bounds on the approximation error.
arXiv Detail & Related papers (2022-07-06T16:31:44Z) - Random matrices in service of ML footprint: ternary random features with
no performance loss [55.30329197651178]
We show that the eigenspectrum of $bf K$ is independent of the distribution of the i.i.d. entries of $bf w$.
We propose a novel random technique, called Ternary Random Feature (TRF)
The computation of the proposed random features requires no multiplication and a factor of $b$ less bits for storage compared to classical random features.
arXiv Detail & Related papers (2021-10-05T09:33:49Z) - Spectral properties of sample covariance matrices arising from random
matrices with independent non identically distributed columns [50.053491972003656]
It was previously shown that the functionals $texttr(AR(z))$, for $R(z) = (frac1nXXT- zI_p)-1$ and $Ain mathcal M_p$ deterministic, have a standard deviation of order $O(|A|_* / sqrt n)$.
Here, we show that $|mathbb E[R(z)] - tilde R(z)|_F
arXiv Detail & Related papers (2021-09-06T14:21:43Z) - Non-PSD Matrix Sketching with Applications to Regression and
Optimization [56.730993511802865]
We present dimensionality reduction methods for non-PSD and square-roots" matrices.
We show how these techniques can be used for multiple downstream tasks.
arXiv Detail & Related papers (2021-06-16T04:07:48Z) - Learning Sparse Graph Laplacian with K Eigenvector Prior via Iterative
GLASSO and Projection [58.5350491065936]
We consider a structural assumption on the graph Laplacian matrix $L$.
The first $K$ eigenvectors of $L$ are pre-selected, e.g., based on domain-specific criteria.
We design an efficient hybrid graphical lasso/projection algorithm to compute the most suitable graph Laplacian matrix $L* in H_u+$ given $barC$.
arXiv Detail & Related papers (2020-10-25T18:12:50Z) - Hutch++: Optimal Stochastic Trace Estimation [75.45968495410048]
We introduce a new randomized algorithm, Hutch++, which computes a $(1 pm epsilon)$ approximation to $tr(A)$ for any positive semidefinite (PSD) $A$.
We show that it significantly outperforms Hutchinson's method in experiments.
arXiv Detail & Related papers (2020-10-19T16:45:37Z) - Signed Graph Metric Learning via Gershgorin Disc Perfect Alignment [46.145969174332485]
We propose a fast general metric learning framework that is entirely projection-free.
We replace the PD cone constraint in the metric learning problem with possible linear constraints per distances.
Experiments show that our graph metric optimization is significantly faster than cone-projection schemes.
arXiv Detail & Related papers (2020-06-15T23:15:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.