Singular value decomposition based matrix surgery
- URL: http://arxiv.org/abs/2302.11446v1
- Date: Wed, 22 Feb 2023 15:30:08 GMT
- Title: Singular value decomposition based matrix surgery
- Authors: Jehan Ghafuri, Sabah Jassim
- Abstract summary: We develop a procedure to reduce and control the condition number of random matrices.
We investigate the effect on the persistent homology (PH) of point clouds of well- and ill-conditioned matrices.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper aims to develop a simple procedure to reduce and control the
condition number of random matrices, and investigate the effect on the
persistent homology (PH) of point clouds of well- and ill-conditioned matrices.
For a square matrix generated randomly using Gaussian/Uniform distribution, the
SVD-Surgery procedure works by: (1) computing its singular value decomposition
(SVD), (2) replacing the diagonal factor by changing a list of the smaller
singular values by a convex linear combination of the entries in the list, and
(3) compute the new matrix by reversing the SVD. Applying SVD-Surgery on a
matrix often results in having different diagonal factor to those of the input
matrix. The spatial distribution of random square matrices are known to be
correlated to the distribution of their condition numbers. The persistent
homology (PH) investigations, therefore, are focused on comparing the effect of
SVD-Surgery on point clouds of large datasets of randomly generated
well-conditioned and ill-conditioned matrices, as well as that of the point
clouds formed by their inverses. This work is motivated by the desire to
stabilise the impact of Deep Learning (DL) training on medical images in terms
of the condition numbers of their sets of convolution filters as a mean of
reducing overfitting and improving robustness against tolerable amounts of
image noise. When applied to convolution filters during training, the
SVD-Surgery acts as a spectral regularisation of the DL model without the need
for learning extra parameters. We shall demonstrate that for several point
clouds of sufficiently large convolution filters our simple strategy preserve
filters norm and reduces the norm of its inverse depending on the chosen linear
combination parameters. Moreover, our approach showed significant improvements
towards the well-conditioning of matrices and stable topological behaviour.
Related papers
- Efficient Adaptation of Pre-trained Vision Transformer via Householder Transformation [53.88562288388169]
A common strategy for.
Efficient Fine-Tuning (PEFT) of pre-trained Vision Transformers (ViTs) involves adapting the model to downstream tasks.
We propose a novel PEFT approach inspired by Singular Value Decomposition (SVD) for representing the adaptation matrix.
SVD decomposes a matrix into the product of a left unitary matrix, a diagonal matrix of scaling values, and a right unitary matrix.
arXiv Detail & Related papers (2024-10-30T12:08:30Z) - Fitting Multilevel Factor Models [41.38783926370621]
We develop a novel, fast implementation of the expectation-maximization algorithm, tailored for multilevel factor models.
We show that the inverse of an invertible PSD MLR matrix is also an MLR matrix with the same sparsity in factors.
We present an algorithm that computes the Cholesky factorization of an expanded matrix with linear time and space complexities.
arXiv Detail & Related papers (2024-09-18T15:39:12Z) - Learning Sparse High-Dimensional Matrix-Valued Graphical Models From Dependent Data [12.94486861344922]
We consider the problem of inferring the conditional independence graph (CIG) of a sparse, high-dimensional, stationary matrix- Gaussian time series.
We consider a sparse-based formulation of the problem with a Kronecker-decomposable power spectral density (PSD)
We illustrate our approach using numerical examples utilizing both synthetic and real data.
arXiv Detail & Related papers (2024-04-29T19:32:50Z) - Large-scale gradient-based training of Mixtures of Factor Analyzers [67.21722742907981]
This article contributes both a theoretical analysis as well as a new method for efficient high-dimensional training by gradient descent.
We prove that MFA training and inference/sampling can be performed based on precision matrices, which does not require matrix inversions after training is completed.
Besides the theoretical analysis and matrices, we apply MFA to typical image datasets such as SVHN and MNIST, and demonstrate the ability to perform sample generation and outlier detection.
arXiv Detail & Related papers (2023-08-26T06:12:33Z) - Memory-Efficient Backpropagation through Large Linear Layers [107.20037639738433]
In modern neural networks like Transformers, linear layers require significant memory to store activations during backward pass.
This study proposes a memory reduction approach to perform backpropagation through linear layers.
arXiv Detail & Related papers (2022-01-31T13:02:41Z) - Robust Low-rank Matrix Completion via an Alternating Manifold Proximal
Gradient Continuation Method [47.80060761046752]
Robust low-rank matrix completion (RMC) has been studied extensively for computer vision, signal processing and machine learning applications.
This problem aims to decompose a partially observed matrix into the superposition of a low-rank matrix and a sparse matrix, where the sparse matrix captures the grossly corrupted entries of the matrix.
A widely used approach to tackle RMC is to consider a convex formulation, which minimizes the nuclear norm of the low-rank matrix (to promote low-rankness) and the l1 norm of the sparse matrix (to promote sparsity).
In this paper, motivated by some recent works on low-
arXiv Detail & Related papers (2020-08-18T04:46:22Z) - Multi-Objective Matrix Normalization for Fine-grained Visual Recognition [153.49014114484424]
Bilinear pooling achieves great success in fine-grained visual recognition (FGVC)
Recent methods have shown that the matrix power normalization can stabilize the second-order information in bilinear features.
We propose an efficient Multi-Objective Matrix Normalization (MOMN) method that can simultaneously normalize a bilinear representation.
arXiv Detail & Related papers (2020-03-30T08:40:35Z) - A Block Coordinate Descent-based Projected Gradient Algorithm for
Orthogonal Non-negative Matrix Factorization [0.0]
This article utilizes the projected gradient method (PG) for a non-negative matrix factorization problem (NMF)
We penalise the orthonormality constraints and apply the PG method via a block coordinate descent approach.
arXiv Detail & Related papers (2020-03-23T13:24:43Z) - Optimal Iterative Sketching with the Subsampled Randomized Hadamard
Transform [64.90148466525754]
We study the performance of iterative sketching for least-squares problems.
We show that the convergence rate for Haar and randomized Hadamard matrices are identical, andally improve upon random projections.
These techniques may be applied to other algorithms that employ randomized dimension reduction.
arXiv Detail & Related papers (2020-02-03T16:17:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.