Complex Network for Complex Problems: A comparative study of CNN and
Complex-valued CNN
- URL: http://arxiv.org/abs/2302.04584v1
- Date: Thu, 9 Feb 2023 11:51:46 GMT
- Title: Complex Network for Complex Problems: A comparative study of CNN and
Complex-valued CNN
- Authors: Soumick Chatterjee, Pavan Tummala, Oliver Speck and Andreas
N\"urnberger
- Abstract summary: Complex-valued convolutional neural networks (CV-CNN) can preserve the algebraic structure of complex-valued input data.
CV-CNNs have double the number of trainable parameters as real-valued CNNs in terms of the actual number of trainable parameters.
This paper presents a comparative study of CNN, CNNx2 (CNN with double the number of trainable parameters as the CNN), and CV-CNN.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural networks, especially convolutional neural networks (CNN), are one of
the most common tools these days used in computer vision. Most of these
networks work with real-valued data using real-valued features. Complex-valued
convolutional neural networks (CV-CNN) can preserve the algebraic structure of
complex-valued input data and have the potential to learn more complex
relationships between the input and the ground-truth. Although some comparisons
of CNNs and CV-CNNs for different tasks have been performed in the past, a
large-scale investigation comparing different models operating on different
tasks has not been conducted. Furthermore, because complex features contain
both real and imaginary components, CV-CNNs have double the number of trainable
parameters as real-valued CNNs in terms of the actual number of trainable
parameters. Whether or not the improvements in performance with CV-CNN observed
in the past have been because of the complex features or just because of having
double the number of trainable parameters has not yet been explored. This paper
presents a comparative study of CNN, CNNx2 (CNN with double the number of
trainable parameters as the CNN), and CV-CNN. The experiments were performed
using seven models for two different tasks - brain tumour classification and
segmentation in brain MRIs. The results have revealed that the CV-CNN models
outperformed the CNN and CNNx2 models.
Related papers
- Model Parallel Training and Transfer Learning for Convolutional Neural Networks by Domain Decomposition [0.0]
Deep convolutional neural networks (CNNs) have been shown to be very successful in a wide range of image processing applications.
Due to their increasing number of model parameters and an increasing availability of large amounts of training data, parallelization strategies to efficiently train complex CNNs are necessary.
arXiv Detail & Related papers (2024-08-26T17:35:01Z) - CNN2GNN: How to Bridge CNN with GNN [59.42117676779735]
We propose a novel CNN2GNN framework to unify CNN and GNN together via distillation.
The performance of distilled boosted'' two-layer GNN on Mini-ImageNet is much higher than CNN containing dozens of layers such as ResNet152.
arXiv Detail & Related papers (2024-04-23T08:19:08Z) - UniHENN: Designing Faster and More Versatile Homomorphic Encryption-based CNNs without im2col [6.496463706588549]
Homomorphic encryption (HE) enables privacy-preserving deep learning by allowing computations on encrypted data without decryption.
deploying convolutional neural networks (CNNs) with HE is challenging due to the need to convert input data into a two-dimensional matrix for convolution using the im2col technique.
UniHENN is a novel HE-based CNN architecture that eliminates the need for im2col, enhancing its versatility and compatibility with a broader range of CNN models.
arXiv Detail & Related papers (2024-02-05T14:52:01Z) - RIC-CNN: Rotation-Invariant Coordinate Convolutional Neural Network [56.42518353373004]
We propose a new convolutional operation, called Rotation-Invariant Coordinate Convolution (RIC-C)
By replacing all standard convolutional layers in a CNN with the corresponding RIC-C, a RIC-CNN can be derived.
It can be observed that RIC-CNN achieves the state-of-the-art classification on the rotated test dataset of MNIST.
arXiv Detail & Related papers (2022-11-21T19:27:02Z) - Towards a General Purpose CNN for Long Range Dependencies in
$\mathrm{N}$D [49.57261544331683]
We propose a single CNN architecture equipped with continuous convolutional kernels for tasks on arbitrary resolution, dimensionality and length without structural changes.
We show the generality of our approach by applying the same CCNN to a wide set of tasks on sequential (1$mathrmD$) and visual data (2$mathrmD$)
Our CCNN performs competitively and often outperforms the current state-of-the-art across all tasks considered.
arXiv Detail & Related papers (2022-06-07T15:48:02Z) - Exploiting Hybrid Models of Tensor-Train Networks for Spoken Command
Recognition [9.262289183808035]
This work aims to design a low complexity spoken command recognition (SCR) system.
We exploit a deep hybrid architecture of a tensor-train (TT) network to build an end-to-end SRC pipeline.
Our proposed CNN+(TT-DNN) model attains a competitive accuracy of 96.31% with 4 times fewer model parameters than the CNN model.
arXiv Detail & Related papers (2022-01-11T05:57:38Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - An Alternative Practice of Tropical Convolution to Traditional
Convolutional Neural Networks [0.5837881923712392]
We propose a new type of CNNs called Tropical Convolutional Neural Networks (TCNNs)
TCNNs are built on tropical convolutions in which the multiplications and additions in conventional convolutional layers are replaced by additions and min/max operations respectively.
We show that TCNN can achieve higher expressive power than ordinary convolutional layers on the MNIST and CIFAR10 image data set.
arXiv Detail & Related papers (2021-03-03T00:13:30Z) - Complex-Valued vs. Real-Valued Neural Networks for Classification
Perspectives: An Example on Non-Circular Data [10.06162739966462]
We show the potential interest of Complex-Valued Neural Network (CVNN) on classification tasks for complex-valued datasets.
CVNN accuracy presents a statistically higher mean and median variance and lower variance than Real-Valued Neural Network (RVNN)
arXiv Detail & Related papers (2020-09-17T14:39:35Z) - Tensor train decompositions on recurrent networks [60.334946204107446]
Matrix product state (MPS) tensor trains have more attractive features than MPOs, in terms of storage reduction and computing time at inference.
We show that MPS tensor trains should be at the forefront of LSTM network compression through a theoretical analysis and practical experiments on NLP task.
arXiv Detail & Related papers (2020-06-09T18:25:39Z) - Neural Additive Models: Interpretable Machine Learning with Neural Nets [77.66871378302774]
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks.
We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models.
NAMs learn a linear combination of neural networks that each attend to a single input feature.
arXiv Detail & Related papers (2020-04-29T01:28:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.