Quantum automated learning with provable and explainable trainability
- URL: http://arxiv.org/abs/2502.05264v1
- Date: Fri, 07 Feb 2025 19:00:02 GMT
- Title: Quantum automated learning with provable and explainable trainability
- Authors: Qi Ye, Shuangyue Geng, Zizhao Han, Weikang Li, L. -M. Duan, Dong-Ling Deng,
- Abstract summary: We introduce quantum automated learning, where no variational parameter is involved and the training process is converted to quantum state preparation.
We show that such a training process can be understood from the perspective of preparing quantum states by imaginary time evolution.
Our results establish an unconventional quantum learning strategy that is gradient-free with provable and explainable trainability.
- Score: 4.305036822025956
- License:
- Abstract: Machine learning is widely believed to be one of the most promising practical applications of quantum computing. Existing quantum machine learning schemes typically employ a quantum-classical hybrid approach that relies crucially on gradients of model parameters. Such an approach lacks provable convergence to global minima and will become infeasible as quantum learning models scale up. Here, we introduce quantum automated learning, where no variational parameter is involved and the training process is converted to quantum state preparation. In particular, we encode training data into unitary operations and iteratively evolve a random initial state under these unitaries and their inverses, with a target-oriented perturbation towards higher prediction accuracy sandwiched in between. Under reasonable assumptions, we rigorously prove that the evolution converges exponentially to the desired state corresponding to the global minimum of the loss function. We show that such a training process can be understood from the perspective of preparing quantum states by imaginary time evolution, where the data-encoded unitaries together with target-oriented perturbations would train the quantum learning model in an automated fashion. We further prove that the quantum automated learning paradigm features good generalization ability with the generalization error upper bounded by the ratio between a logarithmic function of the Hilbert space dimension and the number of training samples. In addition, we carry out extensive numerical simulations on real-life images and quantum data to demonstrate the effectiveness of our approach and validate the assumptions. Our results establish an unconventional quantum learning strategy that is gradient-free with provable and explainable trainability, which would be crucial for large-scale practical applications of quantum computing in machine learning scenarios.
Related papers
- Efficient Learning for Linear Properties of Bounded-Gate Quantum Circuits [63.733312560668274]
Given a quantum circuit containing d tunable RZ gates and G-d Clifford gates, can a learner perform purely classical inference to efficiently predict its linear properties?
We prove that the sample complexity scaling linearly in d is necessary and sufficient to achieve a small prediction error, while the corresponding computational complexity may scale exponentially in d.
We devise a kernel-based learning model capable of trading off prediction error and computational complexity, transitioning from exponential to scaling in many practical settings.
arXiv Detail & Related papers (2024-08-22T08:21:28Z) - The curse of random quantum data [62.24825255497622]
We quantify the performances of quantum machine learning in the landscape of quantum data.
We find that the training efficiency and generalization capabilities in quantum machine learning will be exponentially suppressed with the increase in qubits.
Our findings apply to both the quantum kernel method and the large-width limit of quantum neural networks.
arXiv Detail & Related papers (2024-08-19T12:18:07Z) - Large-scale quantum reservoir learning with an analog quantum computer [45.21335836399935]
We develop a quantum reservoir learning algorithm that harnesses the quantum dynamics of neutral-atom analog quantum computers to process data.
We experimentally implement the algorithm, achieving competitive performance across various categories of machine learning tasks.
Our findings demonstrate the potential of utilizing classically intractable quantum correlations for effective machine learning.
arXiv Detail & Related papers (2024-07-02T18:00:00Z) - Quantum Equilibrium Propagation for efficient training of quantum systems based on Onsager reciprocity [0.0]
Equilibrium propagation (EP) is a procedure that has been introduced and applied to classical energy-based models which relax to an equilibrium.
Here, we show a direct connection between EP and Onsager reciprocity and exploit this to derive a quantum version of EP.
This can be used to optimize loss functions that depend on the expectation values of observables of an arbitrary quantum system.
arXiv Detail & Related papers (2024-06-10T17:22:09Z) - Information-theoretic generalization bounds for learning from quantum data [5.0739329301140845]
We propose a general mathematical formalism for describing quantum learning by training on classical-quantum data.
We prove bounds on the expected generalization error of a quantum learner in terms of classical and quantum information-theoretic quantities.
Our work lays a foundation for a unifying quantum information-theoretic perspective on quantum learning.
arXiv Detail & Related papers (2023-11-09T17:21:38Z) - Quantum data learning for quantum simulations in high-energy physics [55.41644538483948]
We explore the applicability of quantum-data learning to practical problems in high-energy physics.
We make use of ansatz based on quantum convolutional neural networks and numerically show that it is capable of recognizing quantum phases of ground states.
The observation of non-trivial learning properties demonstrated in these benchmarks will motivate further exploration of the quantum-data learning architecture in high-energy physics.
arXiv Detail & Related papers (2023-06-29T18:00:01Z) - Understanding quantum machine learning also requires rethinking
generalization [0.3683202928838613]
We show that traditional approaches to understanding generalization fail to explain the behavior of quantum models.
Experiments reveal that state-of-the-art quantum neural networks accurately fit random states and random labeling of training data.
arXiv Detail & Related papers (2023-06-23T12:04:13Z) - On exploring the potential of quantum auto-encoder for learning quantum systems [60.909817434753315]
We devise three effective QAE-based learning protocols to address three classically computational hard learning problems.
Our work sheds new light on developing advanced quantum learning algorithms to accomplish hard quantum physics and quantum information processing tasks.
arXiv Detail & Related papers (2021-06-29T14:01:40Z) - Quantum-tailored machine-learning characterization of a superconducting
qubit [50.591267188664666]
We develop an approach to characterize the dynamics of a quantum device and learn device parameters.
This approach outperforms physics-agnostic recurrent neural networks trained on numerically generated and experimental data.
This demonstration shows how leveraging domain knowledge improves the accuracy and efficiency of this characterization task.
arXiv Detail & Related papers (2021-06-24T15:58:57Z) - The Hintons in your Neural Network: a Quantum Field Theory View of Deep
Learning [84.33745072274942]
We show how to represent linear and non-linear layers as unitary quantum gates, and interpret the fundamental excitations of the quantum model as particles.
On top of opening a new perspective and techniques for studying neural networks, the quantum formulation is well suited for optical quantum computing.
arXiv Detail & Related papers (2021-03-08T17:24:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.