On establishing learning separations between classical and quantum
machine learning with classical data
- URL: http://arxiv.org/abs/2208.06339v2
- Date: Tue, 4 Jul 2023 12:05:50 GMT
- Title: On establishing learning separations between classical and quantum
machine learning with classical data
- Authors: Casper Gyurik, Vedran Dunjko
- Abstract summary: We discuss the challenges of finding learning problems that quantum learning algorithms can learn much faster than any classical learning algorithm.
We study existing learning problems with a provable quantum speedup to distill sets of more general and sufficient conditions.
These checklists are intended to streamline one's approach to proving quantum speedups for learning problems, or to elucidate bottlenecks.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite years of effort, the quantum machine learning community has only been
able to show quantum learning advantages for certain contrived
cryptography-inspired datasets in the case of classical data. In this note, we
discuss the challenges of finding learning problems that quantum learning
algorithms can learn much faster than any classical learning algorithm, and we
study how to identify such learning problems. Specifically, we reflect on the
main concepts in computational learning theory pertaining to this question, and
we discuss how subtle changes in definitions can mean conceptually
significantly different tasks, which can either lead to a separation or no
separation at all. Moreover, we study existing learning problems with a
provable quantum speedup to distill sets of more general and sufficient
conditions (i.e., ``checklists'') for a learning problem to exhibit a
separation between classical and quantum learners. These checklists are
intended to streamline one's approach to proving quantum speedups for learning
problems, or to elucidate bottlenecks. Finally, to illustrate its application,
we analyze examples of potential separations (i.e., when the learning problem
is build from computational separations, or when the data comes from a quantum
experiment) through the lens of our approach.
Related papers
- A Unified Framework for Neural Computation and Learning Over Time [56.44910327178975]
Hamiltonian Learning is a novel unified framework for learning with neural networks "over time"
It is based on differential equations that: (i) can be integrated without the need of external software solvers; (ii) generalize the well-established notion of gradient-based learning in feed-forward and recurrent networks; (iii) open to novel perspectives.
arXiv Detail & Related papers (2024-09-18T14:57:13Z) - Efficient Learning for Linear Properties of Bounded-Gate Quantum Circuits [63.733312560668274]
Given a quantum circuit containing d tunable RZ gates and G-d Clifford gates, can a learner perform purely classical inference to efficiently predict its linear properties?
We prove that the sample complexity scaling linearly in d is necessary and sufficient to achieve a small prediction error, while the corresponding computational complexity may scale exponentially in d.
We devise a kernel-based learning model capable of trading off prediction error and computational complexity, transitioning from exponential to scaling in many practical settings.
arXiv Detail & Related papers (2024-08-22T08:21:28Z) - Separable Power of Classical and Quantum Learning Protocols Through the Lens of No-Free-Lunch Theorem [70.42372213666553]
The No-Free-Lunch (NFL) theorem quantifies problem- and data-independent generalization errors regardless of the optimization process.
We categorize a diverse array of quantum learning algorithms into three learning protocols designed for learning quantum dynamics under a specified observable.
Our derived NFL theorems demonstrate quadratic reductions in sample complexity across CLC-LPs, ReQu-LPs, and Qu-LPs.
We attribute this performance discrepancy to the unique capacity of quantum-related learning protocols to indirectly utilize information concerning the global phases of non-orthogonal quantum states.
arXiv Detail & Related papers (2024-05-12T09:05:13Z) - Information-theoretic generalization bounds for learning from quantum data [5.0739329301140845]
We propose a general mathematical formalism for describing quantum learning by training on classical-quantum data.
We prove bounds on the expected generalization error of a quantum learner in terms of classical and quantum information-theoretic quantities.
Our work lays a foundation for a unifying quantum information-theoretic perspective on quantum learning.
arXiv Detail & Related papers (2023-11-09T17:21:38Z) - Quantum algorithms: A survey of applications and end-to-end complexities [90.05272647148196]
The anticipated applications of quantum computers span across science and industry.
We present a survey of several potential application areas of quantum algorithms.
We outline the challenges and opportunities in each area in an "end-to-end" fashion.
arXiv Detail & Related papers (2023-10-04T17:53:55Z) - Statistical Complexity of Quantum Learning [32.48879688084909]
This article reviews the complexity of quantum learning using information-theoretic techniques.
We focus on data complexity, copy complexity, and model complexity.
We highlight the differences between quantum and classical learning by addressing both supervised and unsupervised learning.
arXiv Detail & Related papers (2023-09-20T20:04:05Z) - ShadowNet for Data-Centric Quantum System Learning [188.683909185536]
We propose a data-centric learning paradigm combining the strength of neural-network protocols and classical shadows.
Capitalizing on the generalization power of neural networks, this paradigm can be trained offline and excel at predicting previously unseen systems.
We present the instantiation of our paradigm in quantum state tomography and direct fidelity estimation tasks and conduct numerical analysis up to 60 qubits.
arXiv Detail & Related papers (2023-08-22T09:11:53Z) - Exponential separations between classical and quantum learners [0.0]
We discuss how subtle differences in definitions can result in significantly different requirements and tasks for the learner to meet and solve.
We present two new learning separations where the classical difficulty primarily lies in identifying the function generating the data.
arXiv Detail & Related papers (2023-06-28T08:55:56Z) - Classical Verification of Quantum Learning [42.362388367152256]
We develop a framework for classical verification of quantum learning.
We propose a new quantum data access model that we call "mixture-of-superpositions" quantum examples.
Our results demonstrate that the potential power of quantum data for learning tasks, while not unlimited, can be utilized by classical agents.
arXiv Detail & Related papers (2023-06-08T00:31:27Z) - A didactic approach to quantum machine learning with a single qubit [68.8204255655161]
We focus on the case of learning with a single qubit, using data re-uploading techniques.
We implement the different proposed formulations in toy and real-world datasets using the qiskit quantum computing SDK.
arXiv Detail & Related papers (2022-11-23T18:25:32Z) - Power of data in quantum machine learning [2.1012068875084964]
We show that some problems that are classically hard to compute can be easily predicted by classical machines learning from data.
We propose a projected quantum model that provides a simple and rigorous quantum speed-up for a learning problem in the fault-tolerant regime.
arXiv Detail & Related papers (2020-11-03T19:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.