Normalized Gradient Descent for Variational Quantum Algorithms
- URL: http://arxiv.org/abs/2106.10981v1
- Date: Mon, 21 Jun 2021 11:03:12 GMT
- Title: Normalized Gradient Descent for Variational Quantum Algorithms
- Authors: Yudai Suzuki, Hiroshi Yano, Rudy Raymond, Naoki Yamamoto
- Abstract summary: Vari quantum algorithms (VQAs) are promising methods that leverage noisy quantum computers.
NGD method, which employs the normalized gradient vector to update the parameters, has been successfully utilized in several optimization problems.
We propose a new NGD that can attain the faster convergence than the ordinary NGD.
- Score: 4.403985869332685
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational quantum algorithms (VQAs) are promising methods that leverage
noisy quantum computers and classical computing techniques for practical
applications. In VQAs, the classical optimizers such as gradient-based
optimizers are utilized to adjust the parameters of the quantum circuit so that
the objective function is minimized. However, they often suffer from the
so-called vanishing gradient or barren plateau issue. On the other hand, the
normalized gradient descent (NGD) method, which employs the normalized gradient
vector to update the parameters, has been successfully utilized in several
optimization problems. Here, we study the performance of the NGD methods in the
optimization of VQAs for the first time. Our goal is two-fold. The first is to
examine the effectiveness of NGD and its variants for overcoming the vanishing
gradient problems. The second is to propose a new NGD that can attain the
faster convergence than the ordinary NGD. We performed numerical simulations of
these gradient-based optimizers in the context of quantum chemistry where VQAs
are used to find the ground state of a given Hamiltonian. The results show the
effective convergence property of the NGD methods in VQAs, compared to the
relevant optimizers without normalization. Moreover, we make use of some
normalized gradient vectors at the past iteration steps to propose the novel
historical NGD that has a theoretical guarantee to accelerate the convergence
speed, which is observed in the numerical experiments as well.
Related papers
- Application of Langevin Dynamics to Advance the Quantum Natural Gradient Optimization Algorithm [47.47843839099175]
A Quantum Natural Gradient (QNG) algorithm for optimization of variational quantum circuits has been proposed recently.
In this study, we employ the Langevin equation with a QNG force to demonstrate that its discrete-time solution gives a generalized form, which we call Momentum-QNG.
arXiv Detail & Related papers (2024-09-03T15:21:16Z) - Efficient Quantum Gradient and Higher-order Derivative Estimation via Generalized Hadamard Test [2.5545813981422882]
Gradient-based methods are crucial for understanding the behavior of parameterized quantum circuits (PQCs)
Existing gradient estimation methods, such as Finite Difference, Shift Rule, Hadamard Test, and Direct Hadamard Test, often yield suboptimal gradient circuits for certain PQCs.
We introduce the Flexible Hadamard Test, which, when applied to first-order gradient estimation methods, can invert the roles of ansatz generators and observables.
We also introduce Quantum Automatic Differentiation (QAD), a unified gradient method that adaptively selects the best gradient estimation technique for individual parameters within a PQ
arXiv Detail & Related papers (2024-08-10T02:08:54Z) - Qudit inspired optimization for graph coloring [0.0]
We introduce a quantum-inspired algorithm for Graph Coloring Problems (GCPs)
We use qudits in a product state, with each qudit representing a node in the graph and parameterized by d-dimensional spherical coordinates.
We benchmark two optimization strategies: qudit gradient descent (QdGD), initiating qudits in random states and employing gradient descent to minimize a cost function.
arXiv Detail & Related papers (2024-06-02T16:19:55Z) - LAWS: Look Around and Warm-Start Natural Gradient Descent for Quantum
Neural Networks [11.844238544360149]
Vari quantum algorithms (VQAs) have recently received significant attention due to their promising performance in Noisy Intermediate-Scale Quantum computers (NISQ)
VQAs run on parameterized quantum circuits (PQC) with randomlyational parameters are characterized by barren plateaus (BP) where the gradient vanishes exponentially in the number of qubits.
In this paper, we first quantum natural gradient (QNG), which is one of the most popular algorithms used in VQA, from the classical first-order point of optimization.
Then, we proposed a underlineAround underline
arXiv Detail & Related papers (2022-05-05T14:16:40Z) - Adaptive shot allocation for fast convergence in variational quantum
algorithms [0.0]
We present a new gradient descent method using an adaptive number of shots at each step, called the global Coupled Adaptive Number of Shots (gCANS) method.
These improvements reduce both the time and money required to run VQAs on current cloud platforms.
arXiv Detail & Related papers (2021-08-23T22:29:44Z) - Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box
Optimization Framework [100.36569795440889]
This work is on the iteration of zero-th-order (ZO) optimization which does not require first-order information.
We show that with a graceful design in coordinate importance sampling, the proposed ZO optimization method is efficient both in terms of complexity as well as as function query cost.
arXiv Detail & Related papers (2020-12-21T17:29:58Z) - Natural Evolutionary Strategies for Variational Quantum Computation [0.7874708385247353]
Natural evolutionary strategies (NES) are a family of gradient-free black-box optimization algorithms.
This study illustrates their use for the optimization of randomly-d parametrized quantum circuits (PQCs) in the region of vanishing gradients.
arXiv Detail & Related papers (2020-11-30T21:23:38Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z) - IDEAL: Inexact DEcentralized Accelerated Augmented Lagrangian Method [64.15649345392822]
We introduce a framework for designing primal methods under the decentralized optimization setting where local functions are smooth and strongly convex.
Our approach consists of approximately solving a sequence of sub-problems induced by the accelerated augmented Lagrangian method.
When coupled with accelerated gradient descent, our framework yields a novel primal algorithm whose convergence rate is optimal and matched by recently derived lower bounds.
arXiv Detail & Related papers (2020-06-11T18:49:06Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - Towards Better Understanding of Adaptive Gradient Algorithms in
Generative Adversarial Nets [71.05306664267832]
Adaptive algorithms perform gradient updates using the history of gradients and are ubiquitous in training deep neural networks.
In this paper we analyze a variant of OptimisticOA algorithm for nonconcave minmax problems.
Our experiments show that adaptive GAN non-adaptive gradient algorithms can be observed empirically.
arXiv Detail & Related papers (2019-12-26T22:10:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.