Using models to improve optimizers for variational quantum algorithms
- URL: http://arxiv.org/abs/2005.11011v2
- Date: Tue, 11 Aug 2020 18:28:02 GMT
- Title: Using models to improve optimizers for variational quantum algorithms
- Authors: Kevin J. Sung, Jiahao Yao, Matthew P. Harrigan, Nicholas C. Rubin,
Zhang Jiang, Lin Lin, Ryan Babbush, Jarrod R. McClean
- Abstract summary: Variational quantum algorithms are a leading candidate for early applications on noisy intermediate-scale quantum computers.
These algorithms depend on a classical optimization outer-loop that minimizes some function of a parameterized quantum circuit.
We introduce two optimization methods and numerically compare their performance with common methods in use today.
- Score: 1.7475326826331605
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational quantum algorithms are a leading candidate for early applications
on noisy intermediate-scale quantum computers. These algorithms depend on a
classical optimization outer-loop that minimizes some function of a
parameterized quantum circuit. In practice, finite sampling error and gate
errors make this a stochastic optimization with unique challenges that must be
addressed at the level of the optimizer. The sharp trade-off between precision
and sampling time in conjunction with experimental constraints necessitates the
development of new optimization strategies to minimize overall wall clock time
in this setting. In this work, we introduce two optimization methods and
numerically compare their performance with common methods in use today. The
methods are surrogate model-based algorithms designed to improve reuse of
collected data. They do so by utilizing a least-squares quadratic fit of
sampled function values within a moving trusted region to estimate the gradient
or a policy gradient. To make fair comparisons between optimization methods, we
develop experimentally relevant cost models designed to balance efficiency in
testing and accuracy with respect to cloud quantum computing systems. The
results here underscore the need to both use relevant cost models and optimize
hyperparameters of existing optimization methods for competitive performance.
The methods introduced here have several practical advantages in realistic
experimental settings, and we have used one of them successfully in a
separately published experiment on Google's Sycamore device.
Related papers
- Provably Faster Algorithms for Bilevel Optimization via Without-Replacement Sampling [96.47086913559289]
gradient-based algorithms are widely used in bilevel optimization.
We introduce a without-replacement sampling based algorithm which achieves a faster convergence rate.
We validate our algorithms over both synthetic and real-world applications.
arXiv Detail & Related papers (2024-11-07T17:05:31Z) - Model Uncertainty in Evolutionary Optimization and Bayesian Optimization: A Comparative Analysis [5.6787965501364335]
Black-box optimization problems are common in many real-world applications.
These problems require optimization through input-output interactions without access to internal workings.
Two widely used gradient-free optimization techniques are employed to address such challenges.
This paper aims to elucidate the similarities and differences in the utilization of model uncertainty between these two methods.
arXiv Detail & Related papers (2024-03-21T13:59:19Z) - Variational quantum algorithm for enhanced continuous variable optical
phase sensing [0.0]
Variational quantum algorithms (VQAs) are hybrid quantum-classical approaches used for tackling a wide range of problems on noisy quantum devices.
We implement a variational algorithm designed for optimized parameter estimation on a continuous variable platform based on squeezed light.
arXiv Detail & Related papers (2023-12-21T14:11:05Z) - Randomized Benchmarking of Local Zeroth-Order Optimizers for Variational
Quantum Systems [65.268245109828]
We compare the performance of classicals across a series of partially-randomized tasks.
We focus on local zeroth-orders due to their generally favorable performance and query-efficiency on quantum systems.
arXiv Detail & Related papers (2023-10-14T02:13:26Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Surrogate-based optimization for variational quantum algorithms [0.0]
Variational quantum algorithms are a class of techniques intended to be used on near-term quantum computers.
We introduce the idea of learning surrogate models for variational circuits using few experimental measurements.
We then perform parameter optimization using these models as opposed to the original data.
arXiv Detail & Related papers (2022-04-12T00:15:17Z) - Performance comparison of optimization methods on variational quantum
algorithms [2.690135599539986]
Variational quantum algorithms (VQAs) offer a promising path towards using near-term quantum hardware for applications in academic and industrial research.
We study the performance of four commonly used gradient-free optimization methods: SLSQP, COBYLA, CMA-ES, and SPSA.
arXiv Detail & Related papers (2021-11-26T12:13:20Z) - Particle Swarm Optimization: Fundamental Study and its Application to
Optimization and to Jetty Scheduling Problems [0.0]
The advantages of evolutionary algorithms with respect to traditional methods have been greatly discussed in the literature.
While particle swarms share such advantages, they outperform evolutionary algorithms in that they require lower computational cost and easier implementation.
This paper does not intend to study their tuning, general-purpose settings are taken from previous studies, and virtually the same algorithm is used to optimize a variety of notably different problems.
arXiv Detail & Related papers (2021-01-25T02:06:30Z) - Divide and Learn: A Divide and Conquer Approach for Predict+Optimize [50.03608569227359]
The predict+optimize problem combines machine learning ofproblem coefficients with a optimization prob-lem that uses the predicted coefficients.
We show how to directlyexpress the loss of the optimization problem in terms of thepredicted coefficients as a piece-wise linear function.
We propose a novel divide and algorithm to tackle optimization problems without this restriction and predict itscoefficients using the optimization loss.
arXiv Detail & Related papers (2020-12-04T00:26:56Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z) - Cross Entropy Hyperparameter Optimization for Constrained Problem
Hamiltonians Applied to QAOA [68.11912614360878]
Hybrid quantum-classical algorithms such as Quantum Approximate Optimization Algorithm (QAOA) are considered as one of the most encouraging approaches for taking advantage of near-term quantum computers in practical applications.
Such algorithms are usually implemented in a variational form, combining a classical optimization method with a quantum machine to find good solutions to an optimization problem.
In this study we apply a Cross-Entropy method to shape this landscape, which allows the classical parameter to find better parameters more easily and hence results in an improved performance.
arXiv Detail & Related papers (2020-03-11T13:52:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.