Probabilistic Linear Solvers for Machine Learning
- URL: http://arxiv.org/abs/2010.09691v2
- Date: Thu, 22 Oct 2020 19:42:53 GMT
- Title: Probabilistic Linear Solvers for Machine Learning
- Authors: Jonathan Wenger and Philipp Hennig
- Abstract summary: We propose a class of linear solvers which jointly infer the matrix, its inverse and the solution from matrix-vector product observations.
We demonstrate how to incorporate prior spectral information in order to calibrate uncertainty and experimentally showcase the potential of such solvers for machine learning.
- Score: 32.05287257207481
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Linear systems are the bedrock of virtually all numerical computation.
Machine learning poses specific challenges for the solution of such systems due
to their scale, characteristic structure, stochasticity and the central role of
uncertainty in the field. Unifying earlier work we propose a class of
probabilistic linear solvers which jointly infer the matrix, its inverse and
the solution from matrix-vector product observations. This class emerges from a
fundamental set of desiderata which constrains the space of possible algorithms
and recovers the method of conjugate gradients under certain conditions. We
demonstrate how to incorporate prior spectral information in order to calibrate
uncertainty and experimentally showcase the potential of such solvers for
machine learning.
Related papers
- Learning Linear Dynamics from Bilinear Observations [8.238163867581848]
We consider the problem of learning a realization of a partially observed dynamical system with linear state transitions and bilinear observations.
Under very mild assumptions on the process and measurement noises, we provide a finite time analysis for learning the unknown dynamics matrices.
arXiv Detail & Related papers (2024-09-24T23:11:47Z) - Model-Agnostic Zeroth-Order Policy Optimization for Meta-Learning of Ergodic Linear Quadratic Regulators [13.343937277604892]
We study the problem of using meta-learning to deal with uncertainty and heterogeneity in ergodic linear quadratic regulators.
We propose an algorithm that omits the estimation of policy Hessian, which applies to tasks of learning a set of heterogeneous but similar linear dynamic systems.
We provide a convergence result for the exact gradient descent process by analyzing the boundedness and smoothness of the gradient for the meta-objective.
arXiv Detail & Related papers (2024-05-27T17:26:36Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Fairness constraint in Structural Econometrics and Application to fair
estimation using Instrumental Variables [3.265773263570237]
A supervised machine learning algorithm determines a model from a learning sample that will be used to predict new observations.
This information aggregation does not consider any potential selection on unobservables and any status-quo biases which may be contained in the training sample.
The latter bias has raised concerns around the so-called textitfairness of machine learning algorithms, especially towards disadvantaged groups.
arXiv Detail & Related papers (2022-02-16T15:34:07Z) - Quantum algorithms for matrix operations and linear systems of equations [65.62256987706128]
We propose quantum algorithms for matrix operations using the "Sender-Receiver" model.
These quantum protocols can be used as subroutines in other quantum schemes.
arXiv Detail & Related papers (2022-02-10T08:12:20Z) - Joint Learning of Linear Time-Invariant Dynamical Systems [31.879189478584095]
This paper investigates methods for jointly estimating the transition matrices of multiple systems.
We establish finite-time estimation error rates that fully reflect the roles of trajectory lengths, dimension, and number of systems under consideration.
We develop novel techniques that are of interest for addressing similar joint-learning problems.
arXiv Detail & Related papers (2021-12-21T03:09:43Z) - Supervised DKRC with Images for Offline System Identification [77.34726150561087]
Modern dynamical systems are becoming increasingly non-linear and complex.
There is a need for a framework to model these systems in a compact and comprehensive representation for prediction and control.
Our approach learns these basis functions using a supervised learning approach.
arXiv Detail & Related papers (2021-09-06T04:39:06Z) - Online Stochastic Gradient Descent Learns Linear Dynamical Systems from
A Single Trajectory [1.52292571922932]
We show that if the unknown weight matrices describing the system are in Brunovsky canonical form, we can efficiently estimate the ground truth unknown of the system.
Specifically, by deriving concrete bounds, we show that SGD converges linearly in expectation to any arbitrary small Frobenius norm distance from the ground truth weights.
arXiv Detail & Related papers (2021-02-23T17:48:39Z) - Linear embedding of nonlinear dynamical systems and prospects for
efficient quantum algorithms [74.17312533172291]
We describe a method for mapping any finite nonlinear dynamical system to an infinite linear dynamical system (embedding)
We then explore an approach for approximating the resulting infinite linear system with finite linear systems (truncation)
arXiv Detail & Related papers (2020-12-12T00:01:10Z) - General stochastic separation theorems with optimal bounds [68.8204255655161]
Phenomenon of separability was revealed and used in machine learning to correct errors of Artificial Intelligence (AI) systems and analyze AI instabilities.
Errors or clusters of errors can be separated from the rest of the data.
The ability to correct an AI system also opens up the possibility of an attack on it, and the high dimensionality induces vulnerabilities caused by the same separability.
arXiv Detail & Related papers (2020-10-11T13:12:41Z) - Eigendecomposition-Free Training of Deep Networks for Linear
Least-Square Problems [107.3868459697569]
We introduce an eigendecomposition-free approach to training a deep network.
We show that our approach is much more robust than explicit differentiation of the eigendecomposition.
Our method has better convergence properties and yields state-of-the-art results.
arXiv Detail & Related papers (2020-04-15T04:29:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.