A Hybrid Quantum-Classical Approach based on the Hadamard Transform for
the Convolutional Layer
- URL: http://arxiv.org/abs/2305.17510v3
- Date: Thu, 22 Feb 2024 21:38:34 GMT
- Title: A Hybrid Quantum-Classical Approach based on the Hadamard Transform for
the Convolutional Layer
- Authors: Hongyi Pan, Xin Zhu, Salih Atici, Ahmet Enis Cetin
- Abstract summary: We propose a novel Hadamard Transform-based neural network layer for hybrid quantum-classical computing.
The idea is based on the HT convolution theorem which states that the dyadic convolution between two vectors is equivalent to the element-wise multiplication of their HT representation.
- Score: 3.316567107326828
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we propose a novel Hadamard Transform (HT)-based neural
network layer for hybrid quantum-classical computing. It implements the regular
convolutional layers in the Hadamard transform domain. The idea is based on the
HT convolution theorem which states that the dyadic convolution between two
vectors is equivalent to the element-wise multiplication of their HT
representation. Computing the HT is simply the application of a Hadamard gate
to each qubit individually, so the HT computations of our proposed layer can be
implemented on a quantum computer. Compared to the regular Conv2D layer, the
proposed HT-perceptron layer is computationally more efficient. Compared to a
CNN with the same number of trainable parameters and 99.26\% test accuracy, our
HT network reaches 99.31\% test accuracy with 57.1\% MACs reduced in the MNIST
dataset; and in our ImageNet-1K experiments, our HT-based ResNet-50 exceeds the
accuracy of the baseline ResNet-50 by 0.59\% center-crop top-1 accuracy using
11.5\% fewer parameters with 12.6\% fewer MACs.
Related papers
- GMConv: Modulating Effective Receptive Fields for Convolutional Kernels [52.50351140755224]
In convolutional neural networks, the convolutions are performed using a square kernel with a fixed N $times$ N receptive field (RF)
Inspired by the property that ERFs typically exhibit a Gaussian distribution, we propose a Gaussian Mask convolutional kernel (GMConv) in this work.
Our GMConv can directly replace the standard convolutions in existing CNNs and can be easily trained end-to-end by standard back-propagation.
arXiv Detail & Related papers (2023-02-09T10:17:17Z) - DCT Perceptron Layer: A Transform Domain Approach for Convolution Layer [3.506018346865459]
We propose a novel Discrete Cosine Transform (DCT)-based neural network layer which we call DCT-perceptron.
Convolutional filtering operations are performed in the DCT domain using element-wise multiplications.
The DCT-perceptron layer reduces the number of parameters and multiplications significantly.
arXiv Detail & Related papers (2022-11-15T23:44:56Z) - Adaptive Split-Fusion Transformer [90.04885335911729]
We propose an Adaptive Split-Fusion Transformer (ASF-former) to treat convolutional and attention branches differently with adaptive weights.
Experiments on standard benchmarks, such as ImageNet-1K, show that our ASF-former outperforms its CNN, transformer counterparts, and hybrid pilots in terms of accuracy.
arXiv Detail & Related papers (2022-04-26T10:00:28Z) - Block Walsh-Hadamard Transform Based Binary Layers in Deep Neural
Networks [7.906608953906891]
Convolution has been the core operation of modern deep neural networks.
We propose to use binary block Walsh-Hadamard transform (WHT) instead of the Fourier transform.
We use WHT-based binary layers to replace some of the regular convolution layers in deep neural networks.
arXiv Detail & Related papers (2022-01-07T23:52:41Z) - Pruning Neural Networks with Interpolative Decompositions [5.377278489623063]
We introduce a principled approach to neural network pruning that casts the problem as a structured low-rank matrix approximation.
We demonstrate how to prune a neural network by first building a set of primitives to prune a single fully connected or convolution layer.
We achieve an accuracy of 93.62 $pm$ 0.36% using VGG-16 on CIFAR-10, with a 51% FLOPS reduction.
arXiv Detail & Related papers (2021-07-30T20:13:49Z) - Fast Walsh-Hadamard Transform and Smooth-Thresholding Based Binary
Layers in Deep Neural Networks [0.0]
We propose a layer based on fast Walsh-Hadamard transform (WHT) and smooth-thresholding to replace $1times 1$ convolution layers in deep neural networks.
Using these two types of layers, we replace the bottleneck layers in MobileNet-V2 to reduce the network's number of parameters with a slight loss in accuracy.
arXiv Detail & Related papers (2021-04-14T19:23:36Z) - A complete, parallel and autonomous photonic neural network in a
semiconductor multimode laser [0.0]
We show how a fully parallel and fully implemented photonic neural network can be realized using spatially distributed modes of an efficient and fast semiconductor laser.
Importantly, all neural network connections are realized in hardware, and our processor produces results without pre- or post-processing.
We train the readout weights to perform 2-bit header recognition, a 2-bit XOR and 2-bit digital analog conversion, and obtain 0.9 10-3 and 2.9 10-2 error rates for digit recognition and XOR, respectively.
arXiv Detail & Related papers (2020-12-21T07:03:43Z) - DO-Conv: Depthwise Over-parameterized Convolutional Layer [66.46704754669169]
We propose to augment a convolutional layer with an additional depthwise convolution, where each input channel is convolved with a different 2D kernel.
We show with extensive experiments that the mere replacement of conventional convolutional layers with DO-Conv layers boosts the performance of CNNs.
arXiv Detail & Related papers (2020-06-22T06:57:10Z) - Training Binary Neural Networks with Real-to-Binary Convolutions [52.91164959767517]
We show how to train binary networks to within a few percent points of the full precision counterpart.
We show how to build a strong baseline, which already achieves state-of-the-art accuracy.
We show that, when putting all of our improvements together, the proposed model beats the current state of the art by more than 5% top-1 accuracy on ImageNet.
arXiv Detail & Related papers (2020-03-25T17:54:38Z) - Kernel Quantization for Efficient Network Compression [59.55192551370948]
Kernel Quantization (KQ) aims to efficiently convert any pre-trained full-precision convolutional neural network (CNN) model into a low-precision version without significant performance loss.
Inspired by the evolution from weight pruning to filter pruning, we propose to quantize in both kernel and weight level.
Experiments on the ImageNet classification task prove that KQ needs 1.05 and 1.62 bits on average in VGG and ResNet18, respectively, to represent each parameter in the convolution layer.
arXiv Detail & Related papers (2020-03-11T08:00:04Z) - End-to-End Multi-speaker Speech Recognition with Transformer [88.22355110349933]
We replace the RNN-based encoder-decoder in the speech recognition model with a Transformer architecture.
We also modify the self-attention component to be restricted to a segment rather than the whole sequence in order to reduce computation.
arXiv Detail & Related papers (2020-02-10T16:29:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.