Fast and Accurate Pseudoinverse with Sparse Matrix Reordering and
Incremental Approach
- URL: http://arxiv.org/abs/2011.04235v1
- Date: Mon, 9 Nov 2020 07:47:10 GMT
- Title: Fast and Accurate Pseudoinverse with Sparse Matrix Reordering and
Incremental Approach
- Authors: Jinhong Jung and Lee Sael
- Abstract summary: A pseudoinverse is a generalization of a matrix inverse, which has been extensively utilized in machine learning.
FastPI is a novel incremental singular value decomposition (SVD) based pseudoinverse method for sparse matrices.
We show that FastPI computes the pseudoinverse faster than other approximate methods without loss of accuracy.
- Score: 4.710916891482697
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: How can we compute the pseudoinverse of a sparse feature matrix efficiently
and accurately for solving optimization problems? A pseudoinverse is a
generalization of a matrix inverse, which has been extensively utilized as a
fundamental building block for solving linear systems in machine learning.
However, an approximate computation, let alone an exact computation, of
pseudoinverse is very time-consuming due to its demanding time complexity,
which limits it from being applied to large data. In this paper, we propose
FastPI (Fast PseudoInverse), a novel incremental singular value decomposition
(SVD) based pseudoinverse method for sparse matrices. Based on the observation
that many real-world feature matrices are sparse and highly skewed, FastPI
reorders and divides the feature matrix and incrementally computes low-rank SVD
from the divided components. To show the efficacy of proposed FastPI, we apply
them in real-world multi-label linear regression problems. Through extensive
experiments, we demonstrate that FastPI computes the pseudoinverse faster than
other approximate methods without loss of accuracy. %and uses much less memory
compared to full-rank SVD based approach. Results imply that our method
efficiently computes the low-rank pseudoinverse of a large and sparse matrix
that other existing methods cannot handle with limited time and space.
Related papers
- An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - Optimized Sparse Matrix Operations for Reverse Mode Automatic
Differentiation [3.72826300260966]
We present an implementation of a CSR-based sparse matrix wrapper for PyTorch with acceleration for basic matrix operations, as well as automatic differentiability.
We also present several applications of the resulting sparse kernels to optimization problems, demonstrating ease of implementation and performance measurements versus their dense counterparts.
arXiv Detail & Related papers (2022-12-10T00:46:51Z) - RSC: Accelerating Graph Neural Networks Training via Randomized Sparse
Computations [56.59168541623729]
Training graph neural networks (GNNs) is time consuming because sparse graph-based operations are hard to be accelerated by hardware.
We explore trading off the computational precision to reduce the time complexity via sampling-based approximation.
We propose Randomized Sparse Computation, which for the first time demonstrate the potential of training GNNs with approximated operations.
arXiv Detail & Related papers (2022-10-19T17:25:33Z) - High-Dimensional Sparse Bayesian Learning without Covariance Matrices [66.60078365202867]
We introduce a new inference scheme that avoids explicit construction of the covariance matrix.
Our approach couples a little-known diagonal estimation result from numerical linear algebra with the conjugate gradient algorithm.
On several simulations, our method scales better than existing approaches in computation time and memory.
arXiv Detail & Related papers (2022-02-25T16:35:26Z) - Fast Differentiable Matrix Square Root and Inverse Square Root [65.67315418971688]
We propose two more efficient variants to compute the differentiable matrix square root and the inverse square root.
For the forward propagation, one method is to use Matrix Taylor Polynomial (MTP), and the other method is to use Matrix Pad'e Approximants (MPA)
A series of numerical tests show that both methods yield considerable speed-up compared with the SVD or the NS iteration.
arXiv Detail & Related papers (2022-01-29T10:00:35Z) - Fast Differentiable Matrix Square Root [65.67315418971688]
We propose two more efficient variants to compute the differentiable matrix square root.
For the forward propagation, one method is to use Matrix Taylor Polynomial (MTP)
The other method is to use Matrix Pad'e Approximants (MPA)
arXiv Detail & Related papers (2022-01-21T12:18:06Z) - Computationally Efficient Approximations for Matrix-based Renyi's
Entropy [33.72108955447222]
Recently developed matrix based Renyi's entropy enables measurement of information in data.
computation of such quantity involves the trace operator on a PSD matrix $G$ to power $alpha$(i.e., $tr(Galpha)$.
We present computationally efficient approximations to this new entropy functional that can reduce its complexity to even significantly less than $O(n2)$.
arXiv Detail & Related papers (2021-12-27T14:59:52Z) - Learning in High-Dimensional Feature Spaces Using ANOVA-Based Fast
Matrix-Vector Multiplication [0.0]
kernel matrix is typically dense and large-scale. Depending on the dimension of the feature space even the computation of all of its entries in reasonable time becomes a challenging task.
We propose the use of an ANOVA kernel, where we construct several kernels based on lower-dimensional feature spaces for which we provide fast algorithms realizing the matrix-vector products.
Based on a feature grouping approach, we then show how the fast matrix-vector products can be embedded into a learning method choosing kernel ridge regression and the preconditioned conjugate gradient solver.
arXiv Detail & Related papers (2021-11-19T10:29:39Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - Fast Coherent Point Drift [4.369046007546103]
Coherent point drift (CPD) is a classical method for nonrigid point set registration.
By introducing a simple corresponding constraint, we develop a fast implementation of CPD.
Experimental results in 3D point cloud data show that our method can significantly reduce burden of the registration process.
arXiv Detail & Related papers (2020-06-11T09:35:23Z) - Fast Generalized Matrix Regression with Applications in Machine Learning [46.79740387890322]
We propose a fast generalized matrix regression algorithm (Fast GMR) which utilizes sketching technique to solve the GMR problem efficiently.
We apply the Fast GMR algorithm to the symmetric positive definite matrix approximation and single pass singular value decomposition.
arXiv Detail & Related papers (2019-12-27T07:03:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.