Graph Neural Networks for Parameterized Quantum Circuits Expressibility Estimation
- URL: http://arxiv.org/abs/2405.08100v1
- Date: Mon, 13 May 2024 18:26:55 GMT
- Title: Graph Neural Networks for Parameterized Quantum Circuits Expressibility Estimation
- Authors: Shamminuj Aktar, Andreas Bärtschi, Diane Oyen, Stephan Eidenbenz, Abdel-Hameed A. Badawy,
- Abstract summary: This paper introduces a novel approach for expressibility estimation of quantum circuits using Graph Neural Networks (GNNs)
We demonstrate the predictive power of our GNN model with a dataset consisting of 25,000 samples from the noiseless IBM QASM Simulator and 12,000 samples from three distinct noisy quantum backends.
- Score: 5.074765131677166
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Parameterized quantum circuits (PQCs) are fundamental to quantum machine learning (QML), quantum optimization, and variational quantum algorithms (VQAs). The expressibility of PQCs is a measure that determines their capability to harness the full potential of the quantum state space. It is thus a crucial guidepost to know when selecting a particular PQC ansatz. However, the existing technique for expressibility computation through statistical estimation requires a large number of samples, which poses significant challenges due to time and computational resource constraints. This paper introduces a novel approach for expressibility estimation of PQCs using Graph Neural Networks (GNNs). We demonstrate the predictive power of our GNN model with a dataset consisting of 25,000 samples from the noiseless IBM QASM Simulator and 12,000 samples from three distinct noisy quantum backends. The model accurately estimates expressibility, with root mean square errors (RMSE) of 0.05 and 0.06 for the noiseless and noisy backends, respectively. We compare our model's predictions with reference circuits [Sim and others, QuTe'2019] and IBM Qiskit's hardware-efficient ansatz sets to further evaluate our model's performance. Our experimental evaluation in noiseless and noisy scenarios reveals a close alignment with ground truth expressibility values, highlighting the model's efficacy. Moreover, our model exhibits promising extrapolation capabilities, predicting expressibility values with low RMSE for out-of-range qubit circuits trained solely on only up to 5-qubit circuit sets. This work thus provides a reliable means of efficiently evaluating the expressibility of diverse PQCs on noiseless simulators and hardware.
Related papers
- Extending Quantum Perceptrons: Rydberg Devices, Multi-Class Classification, and Error Tolerance [67.77677387243135]
Quantum Neuromorphic Computing (QNC) merges quantum computation with neural computation to create scalable, noise-resilient algorithms for quantum machine learning (QML)
At the core of QNC is the quantum perceptron (QP), which leverages the analog dynamics of interacting qubits to enable universal quantum computation.
arXiv Detail & Related papers (2024-11-13T23:56:20Z) - A Modified Depolarization Approach for Efficient Quantum Machine Learning [0.0]
We propose a modified representation for a single-qubit depolarization channel with two Kraus operators.
Our approach reduces the computational complexity from six to four matrix multiplications per execution of a channel.
This simplified noise model enables more scalable simulations of quantum circuits under depolarization.
arXiv Detail & Related papers (2024-04-10T20:17:40Z) - Probabilistic Sampling of Balanced K-Means using Adiabatic Quantum Computing [93.83016310295804]
AQCs allow to implement problems of research interest, which has sparked the development of quantum representations for computer vision tasks.
In this work, we explore the potential of using this information for probabilistic balanced k-means clustering.
Instead of discarding non-optimal solutions, we propose to use them to compute calibrated posterior probabilities with little additional compute cost.
This allows us to identify ambiguous solutions and data points, which we demonstrate on a D-Wave AQC on synthetic tasks and real visual data.
arXiv Detail & Related papers (2023-10-18T17:59:45Z) - Predicting Expressibility of Parameterized Quantum Circuits using Graph
Neural Network [5.444441239596186]
We propose a novel method based on Graph Neural Networks (GNNs) for predicting the expressibility of Quantum Circuits (PQCs)
By leveraging the graph-based representation of PQCs, our GNN-based model captures intricate relationships between circuit parameters and their resulting expressibility.
Experimental evaluation on a four thousand random PQC dataset and IBM Qiskit's hardware efficient ansatz sets demonstrates the superior performance of our approach.
arXiv Detail & Related papers (2023-09-13T14:08:01Z) - QKSAN: A Quantum Kernel Self-Attention Network [53.96779043113156]
A Quantum Kernel Self-Attention Mechanism (QKSAM) is introduced to combine the data representation merit of Quantum Kernel Methods (QKM) with the efficient information extraction capability of SAM.
A Quantum Kernel Self-Attention Network (QKSAN) framework is proposed based on QKSAM, which ingeniously incorporates the Deferred Measurement Principle (DMP) and conditional measurement techniques.
Four QKSAN sub-models are deployed on PennyLane and IBM Qiskit platforms to perform binary classification on MNIST and Fashion MNIST.
arXiv Detail & Related papers (2023-08-25T15:08:19Z) - Quantum support vector machines for classification and regression on a trapped-ion quantum computer [9.736685719039599]
We examine our quantum machine learning models, which are based on quantum support vector classification (QSVC) and quantum support vector regression (QSVR)
We investigate these models using a quantum-circuit simulator, both with and without noise, as well as the IonQ Harmony quantum processor.
For the classification tasks, the performance of our QSVC models using 4 qubits of the trapped-ion quantum computer was comparable to that obtained from noiseless quantum-circuit simulations.
arXiv Detail & Related papers (2023-07-05T08:06:41Z) - Quantum Circuit Fidelity Improvement with Long Short-Term Memory
Networks [1.2461503242570644]
NISQ computers show great promise in accelerating many tasks that are not practically possible using classical computation.
One important reason is due to the fragile nature of quantum hardware.
As the building blocks of a quantum circuit (QC), quantum gates and qubits are susceptible to external interference.
arXiv Detail & Related papers (2023-03-30T16:44:12Z) - QuEst: Graph Transformer for Quantum Circuit Reliability Estimation [32.89844497610906]
Python library called TorchQuantum can construct, simulate, and train PQC for machine learning tasks.
We propose to leverage a graph transformer model to predict noise impact on circuit fidelity.
Compared with circuit simulators, the predictor has over 200X speedup for estimating the fidelity.
arXiv Detail & Related papers (2022-10-30T02:35:31Z) - Q-ASR: Integer-only Zero-shot Quantization for Efficient Speech
Recognition [65.7040645560855]
We propose Q-ASR, an integer-only, zero-shot quantization scheme for ASR models.
We show negligible WER change as compared to the full-precision baseline models.
Q-ASR exhibits a large compression rate of more than 4x with small WER degradation.
arXiv Detail & Related papers (2021-03-31T06:05:40Z) - Quantum circuit architecture search for variational quantum algorithms [88.71725630554758]
We propose a resource and runtime efficient scheme termed quantum architecture search (QAS)
QAS automatically seeks a near-optimal ansatz to balance benefits and side-effects brought by adding more noisy quantum gates.
We implement QAS on both the numerical simulator and real quantum hardware, via the IBM cloud, to accomplish data classification and quantum chemistry tasks.
arXiv Detail & Related papers (2020-10-20T12:06:27Z) - On the learnability of quantum neural networks [132.1981461292324]
We consider the learnability of the quantum neural network (QNN) built on the variational hybrid quantum-classical scheme.
We show that if a concept can be efficiently learned by QNN, then it can also be effectively learned by QNN even with gate noise.
arXiv Detail & Related papers (2020-07-24T06:34:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.