Bit-bit encoding, optimizer-free training and sub-net initialization: techniques for scalable quantum machine learning
- URL: http://arxiv.org/abs/2501.02148v2
- Date: Wed, 08 Jan 2025 06:12:44 GMT
- Title: Bit-bit encoding, optimizer-free training and sub-net initialization: techniques for scalable quantum machine learning
- Authors: Sonika Johri,
- Abstract summary: We present a quantum classifier that encodes both the input and the output as binary strings.<n>We show that if one parameter is updated at a time, quantum models can be trained in a way that guarantees convergence to a local minimum.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Quantum machine learning for classical data is currently perceived to have a scalability problem due to (i) a bottleneck at the point of loading data into quantum states, (ii) the lack of clarity around good optimization strategies, and (iii) barren plateaus that occur when the model parameters are randomly initialized. In this work, we propose techniques to address all of these issues. First, we present a quantum classifier that encodes both the input and the output as binary strings which results in a model that has no restrictions on expressivity over the encoded data but requires fast classical compression of typical high-dimensional datasets to only the most predictive degrees of freedom. Second, we show that if one parameter is updated at a time, quantum models can be trained without using a classical optimizer in a way that guarantees convergence to a local minimum, something not possible for classical deep learning models. Third, we propose a parameter initialization strategy called sub-net initialization to avoid barren plateaus where smaller models, trained on more compactly encoded data with fewer qubits, are used to initialize models that utilize more qubits. Along with theoretical arguments on efficacy, we demonstrate the combined performance of these methods on subsets of the MNIST dataset for models with an all-to-all connected architecture that use up to 16 qubits in simulation. This allows us to conclude that the loss function consistently decreases as the capability of the model, measured by the number of parameters and qubits, increases, and this behavior is maintained for datasets of varying complexity. Together, these techniques offer a coherent framework for scalable quantum machine learning.
Related papers
- An Efficient Quantum Classifier Based on Hamiltonian Representations [50.467930253994155]
Quantum machine learning (QML) is a discipline that seeks to transfer the advantages of quantum computing to data-driven tasks.
We propose an efficient approach that circumvents the costs associated with data encoding by mapping inputs to a finite set of Pauli strings.
We evaluate our approach on text and image classification tasks, against well-established classical and quantum models.
arXiv Detail & Related papers (2025-04-13T11:49:53Z) - Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - Patch-Based End-to-End Quantum Learning Network for Reduction and Classification of Classical Data [0.22099217573031676]
In the noisy intermediate scale quantum (NISQ) era, the control over the qubits is limited due to errors caused by quantum decoherence, crosstalk, and imperfect calibration.
It is necessary to reduce the size of the large-scale classical data, such as images, when they are to be processed by quantum networks.
In this paper, a dynamic patch-based quantum domain data reduction network with a classical attention mechanism is proposed to avoid such data reductions.
arXiv Detail & Related papers (2024-09-23T16:58:02Z) - Enhancing the performance of Variational Quantum Classifiers with hybrid autoencoders [0.0]
We propose an alternative method which reduces the dimensionality of a given dataset by taking into account the specific quantum embedding that comes after.
This method aspires to make quantum machine learning with VQCs more versatile and effective on datasets of high dimension.
arXiv Detail & Related papers (2024-09-05T08:51:20Z) - $ΞΆ$-QVAE: A Quantum Variational Autoencoder utilizing Regularized Mixed-state Latent Representations [1.0687104237121408]
A major challenge in near-term quantum computing is its application to large real-world datasets due to scarce quantum hardware resources.
We present a fully quantum framework, $zeta$-QVAE, which encompasses all the capabilities of classical VAEs.
Our results consistently indicate that $zeta$-QVAE exhibits similar or better performance compared to matched classical models.
arXiv Detail & Related papers (2024-02-27T18:37:01Z) - Exponential Quantum Communication Advantage in Distributed Inference and Learning [19.827903766111987]
We present a framework for distributed computation over a quantum network.
We show that for models within this framework, inference and training using gradient descent can be performed with exponentially less communication.
We also show that models in this class can encode highly nonlinear features of their inputs, and their expressivity increases exponentially with model depth.
arXiv Detail & Related papers (2023-10-11T02:19:50Z) - NUPES : Non-Uniform Post-Training Quantization via Power Exponent Search [7.971065005161565]
quantization is a technique to convert floating point representations to low bit-width fixed point representations.
We show how to learn new quantized weights over the entire quantized space.
We show the ability of the method to achieve state-of-the-art compression rates in both, data-free and data-driven configurations.
arXiv Detail & Related papers (2023-08-10T14:19:58Z) - A performance characterization of quantum generative models [35.974070202997176]
We compare quantum circuits used for quantum generative modeling.
We learn the underlying probability distribution of the data sets via two popular training methods.
We empirically find that a variant of the discrete architecture, which learns the copula of the probability distribution, outperforms all other methods.
arXiv Detail & Related papers (2023-01-23T11:00:29Z) - A didactic approach to quantum machine learning with a single qubit [68.8204255655161]
We focus on the case of learning with a single qubit, using data re-uploading techniques.
We implement the different proposed formulations in toy and real-world datasets using the qiskit quantum computing SDK.
arXiv Detail & Related papers (2022-11-23T18:25:32Z) - Automatic and effective discovery of quantum kernels [41.61572387137452]
Quantum computing can empower machine learning models by enabling kernel machines to leverage quantum kernels for representing similarity measures between data.<n>We present an approach to this problem, which employs optimization techniques, similar to those used in neural architecture search and AutoML.<n>The results obtained by testing our approach on a high-energy physics problem demonstrate that, in the best-case scenario, we can either match or improve testing accuracy with respect to the manual design approach.
arXiv Detail & Related papers (2022-09-22T16:42:14Z) - ClusterQ: Semantic Feature Distribution Alignment for Data-Free
Quantization [111.12063632743013]
We propose a new and effective data-free quantization method termed ClusterQ.
To obtain high inter-class separability of semantic features, we cluster and align the feature distribution statistics.
We also incorporate the intra-class variance to solve class-wise mode collapse.
arXiv Detail & Related papers (2022-04-30T06:58:56Z) - Towards Quantum Graph Neural Networks: An Ego-Graph Learning Approach [47.19265172105025]
We propose a novel hybrid quantum-classical algorithm for graph-structured data, which we refer to as the Ego-graph based Quantum Graph Neural Network (egoQGNN)
egoQGNN implements the GNN theoretical framework using the tensor product and unity matrix representation, which greatly reduces the number of model parameters required.
The architecture is based on a novel mapping from real-world data to Hilbert space.
arXiv Detail & Related papers (2022-01-13T16:35:45Z) - ECQ$^{\text{x}}$: Explainability-Driven Quantization for Low-Bit and
Sparse DNNs [13.446502051609036]
We develop and describe a novel quantization paradigm for deep neural networks (DNNs)
Our method leverages concepts of explainable AI (XAI) and concepts of information theory.
The ultimate goal is to preserve the most relevant weights in quantization clusters of highest information content.
arXiv Detail & Related papers (2021-09-09T12:57:06Z) - Stabilizing Equilibrium Models by Jacobian Regularization [151.78151873928027]
Deep equilibrium networks (DEQs) are a new class of models that eschews traditional depth in favor of finding the fixed point of a single nonlinear layer.
We propose a regularization scheme for DEQ models that explicitly regularizes the Jacobian of the fixed-point update equations to stabilize the learning of equilibrium models.
We show that this regularization adds only minimal computational cost, significantly stabilizes the fixed-point convergence in both forward and backward passes, and scales well to high-dimensional, realistic domains.
arXiv Detail & Related papers (2021-06-28T00:14:11Z) - GradInit: Learning to Initialize Neural Networks for Stable and
Efficient Training [59.160154997555956]
We present GradInit, an automated and architecture method for initializing neural networks.
It is based on a simple agnostic; the variance of each network layer is adjusted so that a single step of SGD or Adam results in the smallest possible loss value.
It also enables training the original Post-LN Transformer for machine translation without learning rate warmup.
arXiv Detail & Related papers (2021-02-16T11:45:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.