Quantum-Boosted High-Fidelity Deep Learning
- URL: http://arxiv.org/abs/2508.11190v1
- Date: Fri, 15 Aug 2025 03:51:20 GMT
- Title: Quantum-Boosted High-Fidelity Deep Learning
- Authors: Feng-ao Wang, Shaobo Chen, Yao Xuan, Junwei Liu, Qi Gao, Hongdong Zhu, Junjie Hou, Lixin Yuan, Jinyu Cheng, Chenxin Yi, Hai Wei, Yin Ma, Tao Xu, Kai Wen, Yixue Li,
- Abstract summary: We introduce the Quantum Boltzmann Machine-Variational Autoencoder (QBM-VAE), a large-scale and long-time stable hybrid quantum-classical architecture.<n>Our framework leverages a quantum processor for efficient sampling from the Boltzmann distribution, enabling its use as a powerful prior within a deep generative model.
- Score: 7.198071279424711
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: A fundamental limitation of probabilistic deep learning is its predominant reliance on Gaussian priors. This simplistic assumption prevents models from accurately capturing the complex, non-Gaussian landscapes of natural data, particularly in demanding domains like complex biological data, severely hindering the fidelity of the model for scientific discovery. The physically-grounded Boltzmann distribution offers a more expressive alternative, but it is computationally intractable on classical computers. To date, quantum approaches have been hampered by the insufficient qubit scale and operational stability required for the iterative demands of deep learning. Here, we bridge this gap by introducing the Quantum Boltzmann Machine-Variational Autoencoder (QBM-VAE), a large-scale and long-time stable hybrid quantum-classical architecture. Our framework leverages a quantum processor for efficient sampling from the Boltzmann distribution, enabling its use as a powerful prior within a deep generative model. Applied to million-scale single-cell datasets from multiple sources, the QBM-VAE generates a latent space that better preserves complex biological structures, consistently outperforming conventional Gaussian-based deep learning models like VAE and SCVI in essential tasks such as omics data integration, cell-type classification, and trajectory inference. It also provides a typical example of introducing a physics priori into deep learning to drive the model to acquire scientific discovery capabilities that breaks through data limitations. This work provides the demonstration of a practical quantum advantage in deep learning on a large-scale scientific problem and offers a transferable blueprint for developing hybrid quantum AI models.
Related papers
- Quantum AI for Cybersecurity: A hybrid Quantum-Classical models for attack path analysis [1.2676356746752895]
This study investigates the potential of hybrid quantum-classical learning to enhance feature representations for intrusion detection.<n>We show that quantum-enhanced representations demonstrate superior attack recall and improved class separability when data is scarce.<n>Results highlight the potential of quantum embeddings to improve generalization and representation quality in cybersecurity tasks.
arXiv Detail & Related papers (2026-01-05T16:11:39Z) - Hybrid Quantum-Classical Selective State Space Artificial Intelligence [1.4896509623302832]
We propose a Hybrid Quantum Classical selection mechanism for the Mamba architecture for temporal sequence classification problems.<n>Our approach leverages Variational Quantum Circuits (VQCs) as quantum gating modules that both enhance feature extraction and improve suppression of irrelevant information.<n>We analyze how introducing quantum subroutines into large language models (LLMs) impacts their generalization capability, expressivity, and parameter efficiency.
arXiv Detail & Related papers (2025-11-11T15:26:57Z) - VQC-MLPNet: An Unconventional Hybrid Quantum-Classical Architecture for Scalable and Robust Quantum Machine Learning [50.95799256262098]
Variational quantum circuits (VQCs) hold promise for quantum machine learning but face challenges in expressivity, trainability, and noise resilience.<n>We propose VQC-MLPNet, a hybrid architecture where a VQC generates the first-layer weights of a classical multilayer perceptron during training, while inference is performed entirely classically.
arXiv Detail & Related papers (2025-06-12T01:38:15Z) - LatentQGAN: A Hybrid QGAN with Classical Convolutional Autoencoder [7.945302052915863]
A potential application of quantum machine learning is to harness the power of quantum computers for generating classical data.
We propose LatentQGAN, a novel quantum model that uses a hybrid quantum-classical GAN coupled with an autoencoder.
arXiv Detail & Related papers (2024-09-22T23:18:06Z) - Efficient Learning for Linear Properties of Bounded-Gate Quantum Circuits [63.733312560668274]
Given a quantum circuit containing d tunable RZ gates and G-d Clifford gates, can a learner perform purely classical inference to efficiently predict its linear properties?
We prove that the sample complexity scaling linearly in d is necessary and sufficient to achieve a small prediction error, while the corresponding computational complexity may scale exponentially in d.
We devise a kernel-based learning model capable of trading off prediction error and computational complexity, transitioning from exponential to scaling in many practical settings.
arXiv Detail & Related papers (2024-08-22T08:21:28Z) - Neural auto-designer for enhanced quantum kernels [59.616404192966016]
We present a data-driven approach that automates the design of problem-specific quantum feature maps.
Our work highlights the substantial role of deep learning in advancing quantum machine learning.
arXiv Detail & Related papers (2024-01-20T03:11:59Z) - Advances in machine-learning-based sampling motivated by lattice quantum
chromodynamics [4.539861642583362]
This Perspective outlines the advances in ML-based sampling motivated by lattice quantum field theory.
The design of ML algorithms for this application faces profound challenges, including the necessity of scaling custom ML architectures to the largest supercomputers.
If this approach can realize its early promise it will be a transformative step towards first-principles physics calculations in particle, nuclear and condensed matter physics.
arXiv Detail & Related papers (2023-09-03T12:25:59Z) - ShadowNet for Data-Centric Quantum System Learning [188.683909185536]
We propose a data-centric learning paradigm combining the strength of neural-network protocols and classical shadows.
Capitalizing on the generalization power of neural networks, this paradigm can be trained offline and excel at predicting previously unseen systems.
We present the instantiation of our paradigm in quantum state tomography and direct fidelity estimation tasks and conduct numerical analysis up to 60 qubits.
arXiv Detail & Related papers (2023-08-22T09:11:53Z) - A Framework for Demonstrating Practical Quantum Advantage: Racing
Quantum against Classical Generative Models [62.997667081978825]
We build over a proposed framework for evaluating the generalization performance of generative models.
We establish the first comparative race towards practical quantum advantage (PQA) between classical and quantum generative models.
Our results suggest that QCBMs are more efficient in the data-limited regime than the other state-of-the-art classical generative models.
arXiv Detail & Related papers (2023-03-27T22:48:28Z) - The Quantum Path Kernel: a Generalized Quantum Neural Tangent Kernel for
Deep Quantum Machine Learning [52.77024349608834]
Building a quantum analog of classical deep neural networks represents a fundamental challenge in quantum computing.
Key issue is how to address the inherent non-linearity of classical deep learning.
We introduce the Quantum Path Kernel, a formulation of quantum machine learning capable of replicating those aspects of deep machine learning.
arXiv Detail & Related papers (2022-12-22T16:06:24Z) - Generation of High-Resolution Handwritten Digits with an Ion-Trap
Quantum Computer [55.41644538483948]
We implement a quantum-circuit based generative model to learn and sample the prior distribution of a Generative Adversarial Network.
We train this hybrid algorithm on an ion-trap device based on $171$Yb$+$ ion qubits to generate high-quality images.
arXiv Detail & Related papers (2020-12-07T18:51:28Z) - GINNs: Graph-Informed Neural Networks for Multiscale Physics [1.1470070927586016]
Graph-Informed Neural Network (GINN) is a hybrid approach combining deep learning with probabilistic graphical models (PGMs)
GINNs produce kernel density estimates of relevant non-Gaussian, skewed QoIs with tight confidence intervals.
arXiv Detail & Related papers (2020-06-26T05:47:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.