The Bayesian Method of Tensor Networks
- URL: http://arxiv.org/abs/2101.00245v1
- Date: Fri, 1 Jan 2021 14:59:15 GMT
- Title: The Bayesian Method of Tensor Networks
- Authors: Erdong Guo and David Draper
- Abstract summary: We study the Bayesian framework of the Network from two perspective.
We study the Bayesian properties of the Network by visualizing the parameters of the model and the decision boundaries in the two dimensional synthetic data set.
- Score: 1.7894377200944511
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bayesian learning is a powerful learning framework which combines the
external information of the data (background information) with the internal
information (training data) in a logically consistent way in inference and
prediction. By Bayes rule, the external information (prior distribution) and
the internal information (training data likelihood) are combined coherently,
and the posterior distribution and the posterior predictive (marginal)
distribution obtained by Bayes rule summarize the total information needed in
the inference and prediction, respectively. In this paper, we study the
Bayesian framework of the Tensor Network from two perspective. First, we
introduce the prior distribution to the weights in the Tensor Network and
predict the labels of the new observations by the posterior predictive
(marginal) distribution. Since the intractability of the parameter integral in
the normalization constant computation, we approximate the posterior predictive
distribution by Laplace approximation and obtain the out-product approximation
of the hessian matrix of the posterior distribution of the Tensor Network
model. Second, to estimate the parameters of the stationary mode, we propose a
stable initialization trick to accelerate the inference process by which the
Tensor Network can converge to the stationary path more efficiently and stably
with gradient descent method. We verify our work on the MNIST, Phishing Website
and Breast Cancer data set. We study the Bayesian properties of the Bayesian
Tensor Network by visualizing the parameters of the model and the decision
boundaries in the two dimensional synthetic data set. For a application
purpose, our work can reduce the overfitting and improve the performance of
normal Tensor Network model.
Related papers
- Hessian-Free Laplace in Bayesian Deep Learning [44.16006844888796]
Hessian-free Laplace (HFL) approximation uses curvature of both the log posterior and network prediction to estimate its variance.
We show that, under standard assumptions of LA in Bayesian deep learning, HFL targets the same variance as LA, and can be efficiently amortized in a pre-trained network.
arXiv Detail & Related papers (2024-03-15T20:47:39Z) - Learning to solve Bayesian inverse problems: An amortized variational inference approach using Gaussian and Flow guides [0.0]
We develop a methodology that enables real-time inference by learning the Bayesian inverse map, i.e., the map from data to posteriors.
Our approach provides the posterior distribution for a given observation just at the cost of a forward pass of the neural network.
arXiv Detail & Related papers (2023-05-31T16:25:07Z) - Joint Bayesian Inference of Graphical Structure and Parameters with a
Single Generative Flow Network [59.79008107609297]
We propose in this paper to approximate the joint posterior over the structure of a Bayesian Network.
We use a single GFlowNet whose sampling policy follows a two-phase process.
Since the parameters are included in the posterior distribution, this leaves more flexibility for the local probability models.
arXiv Detail & Related papers (2023-05-30T19:16:44Z) - Density Estimation with Autoregressive Bayesian Predictives [1.5771347525430772]
In the context of density estimation, the standard Bayesian approach is to target the posterior predictive.
We develop a novel parameterization of the bandwidth using an autoregressive neural network that maps the data into a latent space.
arXiv Detail & Related papers (2022-06-13T20:43:39Z) - Transformers Can Do Bayesian Inference [56.99390658880008]
We present Prior-Data Fitted Networks (PFNs)
PFNs leverage in-context learning in large-scale machine learning techniques to approximate a large set of posteriors.
We demonstrate that PFNs can near-perfectly mimic Gaussian processes and also enable efficient Bayesian inference for intractable problems.
arXiv Detail & Related papers (2021-12-20T13:07:39Z) - Kalman Bayesian Neural Networks for Closed-form Online Learning [5.220940151628734]
We propose a novel approach for BNN learning via closed-form Bayesian inference.
The calculation of the predictive distribution of the output and the update of the weight distribution are treated as Bayesian filtering and smoothing problems.
This allows closed-form expressions for training the network's parameters in a sequential/online fashion without gradient descent.
arXiv Detail & Related papers (2021-10-03T07:29:57Z) - Sampling-free Variational Inference for Neural Networks with
Multiplicative Activation Noise [51.080620762639434]
We propose a more efficient parameterization of the posterior approximation for sampling-free variational inference.
Our approach yields competitive results for standard regression problems and scales well to large-scale image classification tasks.
arXiv Detail & Related papers (2021-03-15T16:16:18Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.