Is data-efficient learning feasible with quantum models?
- URL: http://arxiv.org/abs/2508.19437v1
- Date: Tue, 26 Aug 2025 21:14:52 GMT
- Title: Is data-efficient learning feasible with quantum models?
- Authors: Alona Sakhnenko, Christian B. Mendl, Jeanette M. Lorenz,
- Abstract summary: We show that quantum kernel methods (QKMs) can achieve low error rates with less training data compared to classical counterparts.<n>We introduce a new analytical tool to the QML domain, derived for classical kernel methods, which can be aimed at investigating the classical-quantum gap.<n>This research contributes to a deeper understanding of the generalization benefits of QKM models and potentially a broader family of QML models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The importance of analyzing nontrivial datasets when testing quantum machine learning (QML) models is becoming increasingly prominent in literature, yet a cohesive framework for understanding dataset characteristics remains elusive. In this work, we concentrate on the size of the dataset as an indicator of its complexity and explores the potential for QML models to demonstrate superior data-efficiency compared to classical models, particularly through the lens of quantum kernel methods (QKMs). We provide a method for generating semi-artificial fully classical datasets, on which we show one of the first evidence of the existence of classical datasets where QKMs require less data during training. Additionally, our study introduces a new analytical tool to the QML domain, derived for classical kernel methods, which can be aimed at investigating the classical-quantum gap. Our empirical results reveal that QKMs can achieve low error rates with less training data compared to classical counterparts. Furthermore, our method allows for the generation of datasets with varying properties, facilitating further investigation into the characteristics of real-world datasets that may be particularly advantageous for QKMs. We also show that the predicted performance from the analytical tool we propose - a generalization metric from classical domain - show great alignment empirical evidence, which fills the gap previously existing in the field. We pave a way to a comprehensive exploration of dataset complexities, providing insights into how these complexities influence QML performance relative to traditional methods. This research contributes to a deeper understanding of the generalization benefits of QKM models and potentially a broader family of QML models, setting the stage for future advancements in the field.
Related papers
- Quantum Extreme Reservoir Computing for Phase Classification of Polymer Alloy Microstructures [0.0]
We apply quantum extreme reservoir computing (QERC) to the classification of polymer alloys generated using self-consistent field theory (SCFT)<n>Results illustrate QERC performance on realistic materials datasets and suggest practical guidelines for quantum encoder design and model generalization.
arXiv Detail & Related papers (2026-01-05T14:25:51Z) - Quantum-Accelerated Neural Imputation with Large Language Models (LLMs) [0.0]
This paper introduces Quantum-UnIMP, a novel framework that integrates shallow quantum circuits into an LLM-based imputation architecture.<n>Our experiments on benchmark mixed-type datasets demonstrate that Quantum-UnIMP reduces imputation error by up to 15.2% for numerical features (RMSE) and improves classification accuracy by 8.7% for categorical features (F1-Score) compared to state-of-the-art classical and LLM-based methods.
arXiv Detail & Related papers (2025-07-11T02:00:06Z) - Entangled mixed-state datasets generation by quantum machine learning [1.5467284361227343]
This paper provides a method for generating mixed-state datasets for entangled-separable classification tasks.<n>It furthers the assembly of quantum entangled datasets, inspires the discovery of new entanglement criteria with both classical and quantum machine learning.
arXiv Detail & Related papers (2025-03-09T05:19:20Z) - Modeling Quantum Machine Learning for Genomic Data Analysis [12.248184406275405]
Quantum Machine Learning (QML) continues to evolve, unlocking new opportunities for diverse applications.<n>We investigate and evaluate the applicability of QML models for binary classification of genome sequence data by employing various feature mapping techniques.<n>We present an open-source, independent Qiskit-based implementation to conduct experiments on a benchmark genomic dataset.
arXiv Detail & Related papers (2025-01-14T15:14:26Z) - Quantum Kernel Learning for Small Dataset Modeling in Semiconductor Fabrication: Application to Ohmic Contact [18.42230728589117]
We develop a quantum kernel-aligned regressor (QKAR) combining a shallow Pauli-Z feature map with a trainable quantum kernel alignment layer.<n>QKAR consistently outperforms classical baselines across multiple metrics.<n>These findings suggest that carefully constructed QML models could provide predictive advantages in data-constrained semiconductor modeling.
arXiv Detail & Related papers (2024-09-17T00:44:49Z) - Benchmarking quantum machine learning kernel training for classification tasks [0.0]
This study focuses on quantum kernel methods in the context of classification tasks.<n>It examines the performance of Quantum Kernel Estimation (QKE) and Quantum Kernel Training (QKT) in connection with two quantum feature mappings.<n> Experimental results indicate that quantum methods exhibit varying performance across different datasets.
arXiv Detail & Related papers (2024-08-17T10:53:06Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - Minimally Supervised Learning using Topological Projections in
Self-Organizing Maps [55.31182147885694]
We introduce a semi-supervised learning approach based on topological projections in self-organizing maps (SOMs)
Our proposed method first trains SOMs on unlabeled data and then a minimal number of available labeled data points are assigned to key best matching units (BMU)
Our results indicate that the proposed minimally supervised model significantly outperforms traditional regression techniques.
arXiv Detail & Related papers (2024-01-12T22:51:48Z) - QKSAN: A Quantum Kernel Self-Attention Network [53.96779043113156]
A Quantum Kernel Self-Attention Mechanism (QKSAM) is introduced to combine the data representation merit of Quantum Kernel Methods (QKM) with the efficient information extraction capability of SAM.
A Quantum Kernel Self-Attention Network (QKSAN) framework is proposed based on QKSAM, which ingeniously incorporates the Deferred Measurement Principle (DMP) and conditional measurement techniques.
Four QKSAN sub-models are deployed on PennyLane and IBM Qiskit platforms to perform binary classification on MNIST and Fashion MNIST.
arXiv Detail & Related papers (2023-08-25T15:08:19Z) - Pre-training Tensor-Train Networks Facilitates Machine Learning with Variational Quantum Circuits [70.97518416003358]
Variational quantum circuits (VQCs) hold promise for quantum machine learning on noisy intermediate-scale quantum (NISQ) devices.
While tensor-train networks (TTNs) can enhance VQC representation and generalization, the resulting hybrid model, TTN-VQC, faces optimization challenges due to the Polyak-Lojasiewicz (PL) condition.
To mitigate this challenge, we introduce Pre+TTN-VQC, a pre-trained TTN model combined with a VQC.
arXiv Detail & Related papers (2023-05-18T03:08:18Z) - A didactic approach to quantum machine learning with a single qubit [68.8204255655161]
We focus on the case of learning with a single qubit, using data re-uploading techniques.
We implement the different proposed formulations in toy and real-world datasets using the qiskit quantum computing SDK.
arXiv Detail & Related papers (2022-11-23T18:25:32Z) - Do Quantum Circuit Born Machines Generalize? [58.720142291102135]
We present the first work in the literature that presents the QCBM's generalization performance as an integral evaluation metric for quantum generative models.
We show that the QCBM is able to effectively learn the reweighted dataset and generate unseen samples with higher quality than those in the training set.
arXiv Detail & Related papers (2022-07-27T17:06:34Z) - Evaluating natural language processing models with generalization
metrics that do not need access to any training or testing data [66.11139091362078]
We provide the first model selection results on large pretrained Transformers from Huggingface using generalization metrics.
Despite their niche status, we find that metrics derived from the heavy-tail (HT) perspective are particularly useful in NLP tasks.
arXiv Detail & Related papers (2022-02-06T20:07:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.