Quantum Gaussian Process Regression for Bayesian Optimization
- URL: http://arxiv.org/abs/2304.12923v1
- Date: Tue, 25 Apr 2023 15:38:19 GMT
- Title: Quantum Gaussian Process Regression for Bayesian Optimization
- Authors: Frederic Rapp and Marco Roth
- Abstract summary: We propose a new approach to Gaussian process regression using quantum kernels based on parameterized quantum circuits.
By employing a hardware-efficient feature map and careful regularization of the Gram matrix, we demonstrate that the variance information of the resulting quantum Gaussian process can be preserved.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Gaussian process regression is a well-established Bayesian machine learning
method. We propose a new approach to Gaussian process regression using quantum
kernels based on parameterized quantum circuits. By employing a
hardware-efficient feature map and careful regularization of the Gram matrix,
we demonstrate that the variance information of the resulting quantum Gaussian
process can be preserved. We also show that quantum Gaussian processes can be
used as a surrogate model for Bayesian optimization, a task that critically
relies on the variance of the surrogate model. To demonstrate the performance
of this quantum Bayesian optimization algorithm, we apply it to the
hyperparameter optimization of a machine learning model which performs
regression on a real-world dataset. We benchmark the quantum Bayesian
optimization against its classical counterpart and show that quantum version
can match its performance.
Related papers
- Advancing quantum process tomography through universal compilation [0.0]
Quantum process tomography (QPT) is crucial for characterizing operations in quantum gates and circuits.
Here, we propose a QPT approach based on universal compilation, which systematically decomposes quantum processes into optimized Kraus operators and Choi matrices.
We benchmark our approach through numerical simulations of random unitary gates, demonstrating highly accurate quantum process characterization.
arXiv Detail & Related papers (2025-04-21T08:34:33Z) - A quantum gradient descent algorithm for optimizing Gaussian Process models [28.16587217223671]
We propose a quantum gradient descent algorithm to optimize the Gaussian Process model.
Our algorithm achieves exponential speedup in computing the gradients of the log marginal likelihood.
arXiv Detail & Related papers (2025-03-22T14:14:31Z) - Sample-efficient Bayesian Optimisation Using Known Invariances [56.34916328814857]
We show that vanilla and constrained BO algorithms are inefficient when optimising invariant objectives.
We derive a bound on the maximum information gain of these invariant kernels.
We use our method to design a current drive system for a nuclear fusion reactor, finding a high-performance solution.
arXiv Detail & Related papers (2024-10-22T12:51:46Z) - Optimizing Unitary Coupled Cluster Wave Functions on Quantum Hardware: Error Bound and Resource-Efficient Optimizer [0.0]
We study the projective quantum eigensolver (PQE) approach to optimizing unitary coupled cluster wave functions on quantum hardware.
The algorithm uses projections of the Schr"odinger equation to efficiently bring the trial state closer to an eigenstate of the Hamiltonian.
We present numerical evidence of superiority over both the optimization introduced in arXiv:2102.00345 and VQE optimized using the Broyden Fletcher Goldfarb Shanno (BFGS) method.
arXiv Detail & Related papers (2024-10-19T15:03:59Z) - Surrogate optimization of variational quantum circuits [1.0546736060336612]
Variational quantum eigensolvers are touted as a near-term algorithm capable of impacting many applications.
Finding algorithms and methods to improve convergence is important to accelerate the capabilities of near-term hardware for VQE.
arXiv Detail & Related papers (2024-04-03T18:00:00Z) - Enhancing Gaussian Process Surrogates for Optimization and Posterior Approximation via Random Exploration [2.984929040246293]
novel noise-free Bayesian optimization strategies that rely on a random exploration step to enhance the accuracy of Gaussian process surrogate models.
New algorithms retain the ease of implementation of the classical GP-UCB, but an additional exploration step facilitates their convergence.
arXiv Detail & Related papers (2024-01-30T14:16:06Z) - Sharp Calibrated Gaussian Processes [58.94710279601622]
State-of-the-art approaches for designing calibrated models rely on inflating the Gaussian process posterior variance.
We present a calibration approach that generates predictive quantiles using a computation inspired by the vanilla Gaussian process posterior variance.
Our approach is shown to yield a calibrated model under reasonable assumptions.
arXiv Detail & Related papers (2023-02-23T12:17:36Z) - Automatic and effective discovery of quantum kernels [43.702574335089736]
Quantum computing can empower machine learning models by enabling kernel machines to leverage quantum kernels for representing similarity measures between data.
We present a different approach, which employs optimization techniques, similar to those used in neural architecture search and AutoML.
The results obtained by testing our approach on a high-energy physics problem demonstrate that, in the best-case scenario, we can either match or improve testing accuracy with respect to the manual design approach.
arXiv Detail & Related papers (2022-09-22T16:42:14Z) - Evaluating the Convergence of Tabu Enhanced Hybrid Quantum Optimization [58.720142291102135]
We introduce the Tabu Enhanced Hybrid Quantum Optimization metaheuristic approach useful for optimization problem solving on a quantum hardware.
We address the theoretical convergence of the proposed scheme from the viewpoint of the collisions in the object which stores the tabu states, based on the Ising model.
arXiv Detail & Related papers (2022-09-05T07:23:03Z) - Surrogate-based optimization for variational quantum algorithms [0.0]
Variational quantum algorithms are a class of techniques intended to be used on near-term quantum computers.
We introduce the idea of learning surrogate models for variational circuits using few experimental measurements.
We then perform parameter optimization using these models as opposed to the original data.
arXiv Detail & Related papers (2022-04-12T00:15:17Z) - Variational Quantum Optimization with Multi-Basis Encodings [62.72309460291971]
We introduce a new variational quantum algorithm that benefits from two innovations: multi-basis graph complexity and nonlinear activation functions.
Our results in increased optimization performance, two increase in effective landscapes and a reduction in measurement progress.
arXiv Detail & Related papers (2021-06-24T20:16:02Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z) - Global Optimization of Gaussian processes [52.77024349608834]
We propose a reduced-space formulation with trained Gaussian processes trained on few data points.
The approach also leads to significantly smaller and computationally cheaper sub solver for lower bounding.
In total, we reduce time convergence by orders of orders of the proposed method.
arXiv Detail & Related papers (2020-05-21T20:59:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.