Absence of Barren Plateaus in Quantum Convolutional Neural Networks
- URL: http://arxiv.org/abs/2011.02966v2
- Date: Mon, 1 Nov 2021 14:27:54 GMT
- Title: Absence of Barren Plateaus in Quantum Convolutional Neural Networks
- Authors: Arthur Pesah, M. Cerezo, Samson Wang, Tyler Volkoff, Andrew T.
Sornborger, Patrick J. Coles
- Abstract summary: Quantum Convolutional Neural Networks (QCNNs) have been proposed.
We rigorously analyze the gradient scaling for the parameters in the QCNN architecture.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Quantum neural networks (QNNs) have generated excitement around the
possibility of efficiently analyzing quantum data. But this excitement has been
tempered by the existence of exponentially vanishing gradients, known as barren
plateau landscapes, for many QNN architectures. Recently, Quantum Convolutional
Neural Networks (QCNNs) have been proposed, involving a sequence of
convolutional and pooling layers that reduce the number of qubits while
preserving information about relevant data features. In this work we rigorously
analyze the gradient scaling for the parameters in the QCNN architecture. We
find that the variance of the gradient vanishes no faster than polynomially,
implying that QCNNs do not exhibit barren plateaus. This provides an analytical
guarantee for the trainability of randomly initialized QCNNs, which highlights
QCNNs as being trainable under random initialization unlike many other QNN
architectures. To derive our results we introduce a novel graph-based method to
analyze expectation values over Haar-distributed unitaries, which will likely
be useful in other contexts. Finally, we perform numerical simulations to
verify our analytical results.
Related papers
- Projected Stochastic Gradient Descent with Quantum Annealed Binary Gradients [51.82488018573326]
We present QP-SBGD, a novel layer-wise optimiser tailored towards training neural networks with binary weights.
BNNs reduce the computational requirements and energy consumption of deep learning models with minimal loss in accuracy.
Our algorithm is implemented layer-wise, making it suitable to train larger networks on resource-limited quantum hardware.
arXiv Detail & Related papers (2023-10-23T17:32:38Z) - Analyzing Convergence in Quantum Neural Networks: Deviations from Neural
Tangent Kernels [20.53302002578558]
A quantum neural network (QNN) is a parameterized mapping efficiently implementable on near-term Noisy Intermediate-Scale Quantum (NISQ) computers.
Despite the existing empirical and theoretical investigations, the convergence of QNN training is not fully understood.
arXiv Detail & Related papers (2023-03-26T22:58:06Z) - QuanGCN: Noise-Adaptive Training for Robust Quantum Graph Convolutional
Networks [124.7972093110732]
We propose quantum graph convolutional networks (QuanGCN), which learns the local message passing among nodes with the sequence of crossing-gate quantum operations.
To mitigate the inherent noises from modern quantum devices, we apply sparse constraint to sparsify the nodes' connections.
Our QuanGCN is functionally comparable or even superior than the classical algorithms on several benchmark graph datasets.
arXiv Detail & Related papers (2022-11-09T21:43:16Z) - GHN-Q: Parameter Prediction for Unseen Quantized Convolutional
Architectures via Graph Hypernetworks [80.29667394618625]
We conduct the first-ever study exploring the use of graph hypernetworks for predicting parameters of unseen quantized CNN architectures.
We focus on a reduced CNN search space and find that GHN-Q can in fact predict quantization-robust parameters for various 8-bit quantized CNNs.
arXiv Detail & Related papers (2022-08-26T08:00:02Z) - Entanglement entropy production in Quantum Neural Networks [0.0]
Quantum Neural Networks (QNN) are considered a candidate for achieving quantum advantage in the Noisy Intermediate Quantum computer (NISQ) era.
We show a universal behavior for the rate at which entanglement is created in any given QNN architecture.
We introduce new measure to characterize the entanglement production in QNNs: the entangling speed.
arXiv Detail & Related papers (2022-06-06T10:17:17Z) - Toward Trainability of Deep Quantum Neural Networks [87.04438831673063]
Quantum Neural Networks (QNNs) with random structures have poor trainability due to the exponentially vanishing gradient as the circuit depth and the qubit number increase.
We provide the first viable solution to the vanishing gradient problem for deep QNNs with theoretical guarantees.
arXiv Detail & Related papers (2021-12-30T10:27:08Z) - Theory of overparametrization in quantum neural networks [0.0]
We rigorously analyze the overparametrization phenomenon in Quantum Neural Networks (QNNs) with periodic structure.
Our results show that the dimension of the Lie algebra obtained from the generators of the QNN is an upper bound for $M_c$.
We then connect the notion of overparametrization to the QNN capacity, so that when a QNN is overparametrized, its capacity achieves its maximum possible value.
arXiv Detail & Related papers (2021-09-23T22:39:48Z) - Toward Trainability of Quantum Neural Networks [87.04438831673063]
Quantum Neural Networks (QNNs) have been proposed as generalizations of classical neural networks to achieve the quantum speed-up.
Serious bottlenecks exist for training QNNs due to the vanishing with gradient rate exponential to the input qubit number.
We show that QNNs with tree tensor and step controlled structures for the application of binary classification. Simulations show faster convergent rates and better accuracy compared to QNNs with random structures.
arXiv Detail & Related papers (2020-11-12T08:32:04Z) - On the learnability of quantum neural networks [132.1981461292324]
We consider the learnability of the quantum neural network (QNN) built on the variational hybrid quantum-classical scheme.
We show that if a concept can be efficiently learned by QNN, then it can also be effectively learned by QNN even with gate noise.
arXiv Detail & Related papers (2020-07-24T06:34:34Z) - Trainability of Dissipative Perceptron-Based Quantum Neural Networks [0.8258451067861933]
We analyze the gradient scaling (and hence the trainability) for a recently proposed architecture that we called dissipative QNNs (DQNNs)
We find that DQNNs can exhibit barren plateaus, i.e., gradients that vanish exponentially in the number of qubits.
arXiv Detail & Related papers (2020-05-26T00:59:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.