Impact of Data Augmentation on QCNNs
- URL: http://arxiv.org/abs/2312.00358v1
- Date: Fri, 1 Dec 2023 05:28:19 GMT
- Title: Impact of Data Augmentation on QCNNs
- Authors: Leting Zhouli, Peiyong Wang, Udaya Parampalli
- Abstract summary: Quantum Convolutional Neural Networks (QCNNs) are proposed as a novel generalization to CNNs by using quantum mechanisms.
This paper implements and compares both CNNs and QCNNs by testing losses and prediction accuracy on three commonly used datasets.
- Score: 1.1510009152620664
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, Classical Convolutional Neural Networks (CNNs) have been
applied for image recognition successfully. Quantum Convolutional Neural
Networks (QCNNs) are proposed as a novel generalization to CNNs by using
quantum mechanisms. The quantum mechanisms lead to an efficient training
process in QCNNs by reducing the size of input from $N$ to $log_2N$. This paper
implements and compares both CNNs and QCNNs by testing losses and prediction
accuracy on three commonly used datasets. The datasets include the MNIST
hand-written digits, Fashion MNIST and cat/dog face images. Additionally, data
augmentation (DA), a technique commonly used in CNNs to improve the performance
of classification by generating similar images based on original inputs, is
also implemented in QCNNs. Surprisingly, the results showed that data
augmentation didn't improve QCNNs performance. The reasons and logic behind
this result are discussed, hoping to expand our understanding of Quantum
machine learning theory.
Related papers
- CNN2GNN: How to Bridge CNN with GNN [59.42117676779735]
We propose a novel CNN2GNN framework to unify CNN and GNN together via distillation.
The performance of distilled boosted'' two-layer GNN on Mini-ImageNet is much higher than CNN containing dozens of layers such as ResNet152.
arXiv Detail & Related papers (2024-04-23T08:19:08Z) - Correlated Optical Convolutional Neural Network with Quantum Speedup [9.194748023567886]
We show that the correlated optical convolutional neural network (COCNN) can exhibit quantum speedup in the training process.
Our proposal opens up a new avenue for realizing the ONNs with the quantum speedup.
arXiv Detail & Related papers (2024-02-01T11:17:09Z) - A Quantum Convolutional Neural Network Approach for Object Detection and
Classification [0.0]
The time and accuracy of QCNNs are compared with classical CNNs and ANN models under different conditions.
The analysis shows that QCNNs have the potential to outperform both classical CNNs and ANN models in terms of accuracy and efficiency for certain applications.
arXiv Detail & Related papers (2023-07-17T02:38:04Z) - Predict better with less training data using a QNN [1.7481852615249125]
We describe a quanvolutional neural network (QNN) algorithm that efficiently maps classical image data to quantum states.
We empirically observe a genuine quantum advantage for an industrial application where the advantage is due to superior data encoding.
arXiv Detail & Related papers (2022-06-08T15:25:58Z) - Quantum convolutional neural network for classical data classification [0.8057006406834467]
We benchmark fully parameterized quantum convolutional neural networks (QCNNs) for classical data classification.
We propose a quantum neural network model inspired by CNN that only uses two-qubit interactions throughout the entire algorithm.
arXiv Detail & Related papers (2021-08-02T06:48:34Z) - The dilemma of quantum neural networks [63.82713636522488]
We show that quantum neural networks (QNNs) fail to provide any benefit over classical learning models.
QNNs suffer from the severely limited effective model capacity, which incurs poor generalization on real-world datasets.
These results force us to rethink the role of current QNNs and to design novel protocols for solving real-world problems with quantum advantages.
arXiv Detail & Related papers (2021-06-09T10:41:47Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - Toward Trainability of Quantum Neural Networks [87.04438831673063]
Quantum Neural Networks (QNNs) have been proposed as generalizations of classical neural networks to achieve the quantum speed-up.
Serious bottlenecks exist for training QNNs due to the vanishing with gradient rate exponential to the input qubit number.
We show that QNNs with tree tensor and step controlled structures for the application of binary classification. Simulations show faster convergent rates and better accuracy compared to QNNs with random structures.
arXiv Detail & Related papers (2020-11-12T08:32:04Z) - A Tutorial on Quantum Convolutional Neural Networks (QCNN) [11.79760591464748]
Convolutional Neural Network (CNN) is a popular model in computer vision.
CNN is challenging to learn efficiently if the given dimension of data or model becomes too large.
Quantum Convolutional Neural Network (QCNN) provides a new solution to a problem to solve with CNN using a quantum computing environment.
arXiv Detail & Related papers (2020-09-20T12:29:05Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.