Bit-bit encoding, optimizer-free training and sub-net initialization: techniques for scalable quantum machine learning
- URL: http://arxiv.org/abs/2501.02148v2
- Date: Wed, 08 Jan 2025 06:12:44 GMT
- Title: Bit-bit encoding, optimizer-free training and sub-net initialization: techniques for scalable quantum machine learning
- Authors: Sonika Johri,
- Abstract summary: We present a quantum classifier that encodes both the input and the output as binary strings.
We show that if one parameter is updated at a time, quantum models can be trained in a way that guarantees convergence to a local minimum.
- Score: 0.0
- License:
- Abstract: Quantum machine learning for classical data is currently perceived to have a scalability problem due to (i) a bottleneck at the point of loading data into quantum states, (ii) the lack of clarity around good optimization strategies, and (iii) barren plateaus that occur when the model parameters are randomly initialized. In this work, we propose techniques to address all of these issues. First, we present a quantum classifier that encodes both the input and the output as binary strings which results in a model that has no restrictions on expressivity over the encoded data but requires fast classical compression of typical high-dimensional datasets to only the most predictive degrees of freedom. Second, we show that if one parameter is updated at a time, quantum models can be trained without using a classical optimizer in a way that guarantees convergence to a local minimum, something not possible for classical deep learning models. Third, we propose a parameter initialization strategy called sub-net initialization to avoid barren plateaus where smaller models, trained on more compactly encoded data with fewer qubits, are used to initialize models that utilize more qubits. Along with theoretical arguments on efficacy, we demonstrate the combined performance of these methods on subsets of the MNIST dataset for models with an all-to-all connected architecture that use up to 16 qubits in simulation. This allows us to conclude that the loss function consistently decreases as the capability of the model, measured by the number of parameters and qubits, increases, and this behavior is maintained for datasets of varying complexity. Together, these techniques offer a coherent framework for scalable quantum machine learning.
Related papers
- Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - Patch-Based End-to-End Quantum Learning Network for Reduction and Classification of Classical Data [0.22099217573031676]
In the noisy intermediate scale quantum (NISQ) era, the control over the qubits is limited due to errors caused by quantum decoherence, crosstalk, and imperfect calibration.
It is necessary to reduce the size of the large-scale classical data, such as images, when they are to be processed by quantum networks.
In this paper, a dynamic patch-based quantum domain data reduction network with a classical attention mechanism is proposed to avoid such data reductions.
arXiv Detail & Related papers (2024-09-23T16:58:02Z) - Enhancing the performance of Variational Quantum Classifiers with hybrid autoencoders [0.0]
We propose an alternative method which reduces the dimensionality of a given dataset by taking into account the specific quantum embedding that comes after.
This method aspires to make quantum machine learning with VQCs more versatile and effective on datasets of high dimension.
arXiv Detail & Related papers (2024-09-05T08:51:20Z) - $ΞΆ$-QVAE: A Quantum Variational Autoencoder utilizing Regularized Mixed-state Latent Representations [1.0687104237121408]
A major challenge in near-term quantum computing is its application to large real-world datasets due to scarce quantum hardware resources.
We present a fully quantum framework, $zeta$-QVAE, which encompasses all the capabilities of classical VAEs.
Our results consistently indicate that $zeta$-QVAE exhibits similar or better performance compared to matched classical models.
arXiv Detail & Related papers (2024-02-27T18:37:01Z) - NUPES : Non-Uniform Post-Training Quantization via Power Exponent Search [7.971065005161565]
quantization is a technique to convert floating point representations to low bit-width fixed point representations.
We show how to learn new quantized weights over the entire quantized space.
We show the ability of the method to achieve state-of-the-art compression rates in both, data-free and data-driven configurations.
arXiv Detail & Related papers (2023-08-10T14:19:58Z) - A didactic approach to quantum machine learning with a single qubit [68.8204255655161]
We focus on the case of learning with a single qubit, using data re-uploading techniques.
We implement the different proposed formulations in toy and real-world datasets using the qiskit quantum computing SDK.
arXiv Detail & Related papers (2022-11-23T18:25:32Z) - ClusterQ: Semantic Feature Distribution Alignment for Data-Free
Quantization [111.12063632743013]
We propose a new and effective data-free quantization method termed ClusterQ.
To obtain high inter-class separability of semantic features, we cluster and align the feature distribution statistics.
We also incorporate the intra-class variance to solve class-wise mode collapse.
arXiv Detail & Related papers (2022-04-30T06:58:56Z) - ECQ$^{\text{x}}$: Explainability-Driven Quantization for Low-Bit and
Sparse DNNs [13.446502051609036]
We develop and describe a novel quantization paradigm for deep neural networks (DNNs)
Our method leverages concepts of explainable AI (XAI) and concepts of information theory.
The ultimate goal is to preserve the most relevant weights in quantization clusters of highest information content.
arXiv Detail & Related papers (2021-09-09T12:57:06Z) - Stabilizing Equilibrium Models by Jacobian Regularization [151.78151873928027]
Deep equilibrium networks (DEQs) are a new class of models that eschews traditional depth in favor of finding the fixed point of a single nonlinear layer.
We propose a regularization scheme for DEQ models that explicitly regularizes the Jacobian of the fixed-point update equations to stabilize the learning of equilibrium models.
We show that this regularization adds only minimal computational cost, significantly stabilizes the fixed-point convergence in both forward and backward passes, and scales well to high-dimensional, realistic domains.
arXiv Detail & Related papers (2021-06-28T00:14:11Z) - GradInit: Learning to Initialize Neural Networks for Stable and
Efficient Training [59.160154997555956]
We present GradInit, an automated and architecture method for initializing neural networks.
It is based on a simple agnostic; the variance of each network layer is adjusted so that a single step of SGD or Adam results in the smallest possible loss value.
It also enables training the original Post-LN Transformer for machine translation without learning rate warmup.
arXiv Detail & Related papers (2021-02-16T11:45:35Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.