Adaptive shot allocation for fast convergence in variational quantum
algorithms
- URL: http://arxiv.org/abs/2108.10434v1
- Date: Mon, 23 Aug 2021 22:29:44 GMT
- Title: Adaptive shot allocation for fast convergence in variational quantum
algorithms
- Authors: Andi Gu, Angus Lowe, Pavel A. Dub, Patrick J. Coles, Andrew Arrasmith
- Abstract summary: We present a new gradient descent method using an adaptive number of shots at each step, called the global Coupled Adaptive Number of Shots (gCANS) method.
These improvements reduce both the time and money required to run VQAs on current cloud platforms.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational Quantum Algorithms (VQAs) are a promising approach for practical
applications like chemistry and materials science on near-term quantum
computers as they typically reduce quantum resource requirements. However, in
order to implement VQAs, an efficient classical optimization strategy is
required. Here we present a new stochastic gradient descent method using an
adaptive number of shots at each step, called the global Coupled Adaptive
Number of Shots (gCANS) method, which improves on prior art in both the number
of iterations as well as the number of shots required. These improvements
reduce both the time and money required to run VQAs on current cloud platforms.
We analytically prove that in a convex setting gCANS achieves geometric
convergence to the optimum. Further, we numerically investigate the performance
of gCANS on some chemical configuration problems. We also consider finding the
ground state for an Ising model with different numbers of spins to examine the
scaling of the method. We find that for these problems, gCANS compares
favorably to all of the other optimizers we consider.
Related papers
- Bayesian Optimization Priors for Efficient Variational Quantum Algorithms [0.0]
Quantum computers currently rely on a quantum-classical approach known as Variational Quantum Algorithms (VQAs) to solve problems.
We propose a hybrid framework for basic computational optimization that reduces the number of shots per time is charged.
Using both proposed features, we show that using both proposed features statistically outperforms an implementation within VQAs for simulations.
arXiv Detail & Related papers (2024-06-20T18:00:12Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Iteration Complexity of Variational Quantum Algorithms [5.203200173190989]
We argue that noise makes evaluations of the objective function via quantum circuits biased.
We derive the missing guarantees and find that the rate of convergence is unaffected.
arXiv Detail & Related papers (2022-09-21T19:18:41Z) - A Gauss-Newton based Quantum Algorithm for Combinatorial Optimization [0.0]
We present a Gauss-Newton based quantum algorithm (GNQA) for optimization problems that, under optimal conditions, rapidly converge towards one of the optimal solutions without being trapped in local minima or plateaus.
Our approach mitigates those by employing a tensor product state that accurately represents the optimal solution, and an appropriate function for the Hamiltonian, containing all the combinations of binary variables.
Numerical experiments presented here demonstrate the effectiveness of our approach, and they show that GNQA outperforms other optimization methods in both convergence properties and accuracy for all problems considered here.
arXiv Detail & Related papers (2022-03-25T23:49:31Z) - Faster One-Sample Stochastic Conditional Gradient Method for Composite
Convex Minimization [61.26619639722804]
We propose a conditional gradient method (CGM) for minimizing convex finite-sum objectives formed as a sum of smooth and non-smooth terms.
The proposed method, equipped with an average gradient (SAG) estimator, requires only one sample per iteration. Nevertheless, it guarantees fast convergence rates on par with more sophisticated variance reduction techniques.
arXiv Detail & Related papers (2022-02-26T19:10:48Z) - SUPER-ADAM: Faster and Universal Framework of Adaptive Gradients [99.13839450032408]
It is desired to design a universal framework for adaptive algorithms to solve general problems.
In particular, our novel framework provides adaptive methods under non convergence support for setting.
arXiv Detail & Related papers (2021-06-15T15:16:28Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - An adaptive quantum approximate optimization algorithm for solving
combinatorial problems on a quantum computer [0.0]
The quantum approximate optimization algorithm (QAOA) is a hybrid variational quantum-classical algorithm that solves optimization problems.
We develop an iterative version of QAOA that is problem-tailored, and which can also be adapted to specific hardware constraints.
We simulate the algorithm on a class of Max-Cut graph problems and show that it converges much faster than the standard QAOA.
arXiv Detail & Related papers (2020-05-20T18:00:01Z) - Operator Sampling for Shot-frugal Optimization in Variational Algorithms [0.860968116787044]
We introduce a strategy for reducing the number of measurements (i.e., shots) by randomly sampling operators $h_i$ from the overall $H = sum_i c_i h_i$.
We implement this and other Samplings to find the ground states of molecules H$$, LiH, and BeH$, without and with quantum hardware noise, and Rosalin outperforms other Samplings in most cases.
arXiv Detail & Related papers (2020-04-14T01:00:07Z) - Cross Entropy Hyperparameter Optimization for Constrained Problem
Hamiltonians Applied to QAOA [68.11912614360878]
Hybrid quantum-classical algorithms such as Quantum Approximate Optimization Algorithm (QAOA) are considered as one of the most encouraging approaches for taking advantage of near-term quantum computers in practical applications.
Such algorithms are usually implemented in a variational form, combining a classical optimization method with a quantum machine to find good solutions to an optimization problem.
In this study we apply a Cross-Entropy method to shape this landscape, which allows the classical parameter to find better parameters more easily and hence results in an improved performance.
arXiv Detail & Related papers (2020-03-11T13:52:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.