Training Gaussian Boson Sampling Distributions
- URL: http://arxiv.org/abs/2004.04770v1
- Date: Thu, 9 Apr 2020 18:53:55 GMT
- Title: Training Gaussian Boson Sampling Distributions
- Authors: Leonardo Banchi, Nicol\'as Quesada, Juan Miguel Arrazola
- Abstract summary: We derive analytical gradient formulas for the GBS distribution, which can be used to train devices using standard methods.
In the case of training using a Kullback-Leibler divergence or log-likelihood cost function, we show that gradients can be computed classically, leading to fast training.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gaussian Boson Sampling (GBS) is a near-term platform for photonic quantum
computing. Applications have been developed which rely on directly programming
GBS devices, but the ability to train and optimize circuits has been a key
missing ingredient for developing new algorithms. In this work, we derive
analytical gradient formulas for the GBS distribution, which can be used to
train devices using standard methods based on gradient descent. We introduce a
parametrization of the distribution that allows the gradient to be estimated by
sampling from the same device that is being optimized. In the case of training
using a Kullback-Leibler divergence or log-likelihood cost function, we show
that gradients can be computed classically, leading to fast training. We
illustrate these results with numerical experiments in stochastic optimization
and unsupervised learning. As a particular example, we introduce the
variational Ising solver, a hybrid algorithm for training GBS devices to sample
ground states of a classical Ising model with high probability.
Related papers
- Stochastic Gradient Descent for Gaussian Processes Done Right [86.83678041846971]
We show that when emphdone right -- by which we mean using specific insights from optimisation and kernel communities -- gradient descent is highly effective.
We introduce a emphstochastic dual descent algorithm, explain its design in an intuitive manner and illustrate the design choices.
Our method places Gaussian process regression on par with state-of-the-art graph neural networks for molecular binding affinity prediction.
arXiv Detail & Related papers (2023-10-31T16:15:13Z) - Sampling from Gaussian Process Posteriors using Stochastic Gradient
Descent [43.097493761380186]
gradient algorithms are an efficient method of approximately solving linear systems.
We show that gradient descent produces accurate predictions, even in cases where it does not converge quickly to the optimum.
Experimentally, gradient descent achieves state-of-the-art performance on sufficiently large-scale or ill-conditioned regression tasks.
arXiv Detail & Related papers (2023-06-20T15:07:37Z) - Stochastic Marginal Likelihood Gradients using Neural Tangent Kernels [78.6096486885658]
We introduce lower bounds to the linearized Laplace approximation of the marginal likelihood.
These bounds are amenable togradient-based optimization and allow to trade off estimation accuracy against computational complexity.
arXiv Detail & Related papers (2023-06-06T19:02:57Z) - Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model [89.8764435351222]
We propose a new family of unbiased estimators called WTA-CRS, for matrix production with reduced variance.
Our work provides both theoretical and experimental evidence that, in the context of tuning transformers, our proposed estimators exhibit lower variance compared to existing ones.
arXiv Detail & Related papers (2023-05-24T15:52:08Z) - Federated Optimization Algorithms with Random Reshuffling and Gradient
Compression [2.7554288121906296]
We provide the first analysis of methods with gradient compression and without-replacement sampling.
We show how to reduce the variance coming from gradient quantization through the use of control iterates.
We outline several settings in which they improve upon existing algorithms.
arXiv Detail & Related papers (2022-06-14T17:36:47Z) - Protocols for Trainable and Differentiable Quantum Generative Modelling [21.24186888129542]
We propose an approach for learning probability distributions as differentiable quantum circuits (DQC)
We perform training of a DQC-based model, where data is encoded in a latent space with a phase feature map, followed by a variational quantum circuit.
This allows fast sampling from parametrized distributions using a single-shot readout.
arXiv Detail & Related papers (2022-02-16T18:55:48Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Efficient calculation of gradients in classical simulations of
variational quantum algorithms [0.0]
We present a novel derivation of an emulation strategy to precisely calculate the gradient in O(P) time.
Our strategy is very simple, uses only 'apply gate', 'clone state' and 'inner product' primitives.
It is compatible with gate parallelisation schemes, and hardware accelerated and distributed simulators.
arXiv Detail & Related papers (2020-09-06T21:39:44Z) - Towards Better Understanding of Adaptive Gradient Algorithms in
Generative Adversarial Nets [71.05306664267832]
Adaptive algorithms perform gradient updates using the history of gradients and are ubiquitous in training deep neural networks.
In this paper we analyze a variant of OptimisticOA algorithm for nonconcave minmax problems.
Our experiments show that adaptive GAN non-adaptive gradient algorithms can be observed empirically.
arXiv Detail & Related papers (2019-12-26T22:10:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.