Complex-valued Neural Networks -- Theory and Analysis
- URL: http://arxiv.org/abs/2312.06087v1
- Date: Mon, 11 Dec 2023 03:24:26 GMT
- Title: Complex-valued Neural Networks -- Theory and Analysis
- Authors: Rayyan Abdalla
- Abstract summary: This work addresses different structures and classification of CVNNs.
The theory behind complex activation functions, implications related to complex differentiability and special activations for CVNN output layers are presented.
The objective of this work is to understand the dynamics and most recent developments of CVNNs.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Complex-valued neural networks (CVNNs) have recently been successful in
various pioneering areas which involve wave-typed information and
frequency-domain processing. This work addresses different structures and
classification of CVNNs. The theory behind complex activation functions,
implications related to complex differentiability and special activations for
CVNN output layers are presented. The work also discusses CVNN learning and
optimization using gradient and non-gradient based algorithms. Complex
Backpropagation utilizing complex chain rule is also explained in terms of
Wirtinger calculus. Moreover, special modules for building CVNN models, such as
complex batch normalization and complex random initialization are also
discussed. The work also highlights libraries and software blocks proposed for
CVNN implementations and discusses future directions. The objective of this
work is to understand the dynamics and most recent developments of CVNNs.
Related papers
- Comprehensive Survey of Complex-Valued Neural Networks: Insights into Backpropagation and Activation Functions [0.0]
Despite the prevailing use of real-number implementations in current ANN frameworks, there is a growing interest in developing ANNs that utilize complex numbers.
This paper presents a survey of recent advancements in complex-valued neural networks (CVNNs)
We delve into the extension of the backpropagation algorithm to the complex domain, which enables the training of neural networks with complex-valued inputs, weights, AFs, and outputs.
arXiv Detail & Related papers (2024-07-27T13:47:16Z) - On the Computational Complexities of Complex-valued Neural Networks [0.0]
Complex-valued neural networks (CVNNs) are nonlinear filters used in the digital signal processing of complex-domain data.
This paper presents both the quantitative and computational complexities of CVNNs.
arXiv Detail & Related papers (2023-10-19T18:14:04Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Theory and Implementation of Complex-Valued Neural Networks [9.6556424340252]
This work explains in detail the theory behind Complex-Valued Neural Network (CVNN)
It includes Wirtinger calculus, complex backpropagation, and basic modules such as complex layers.
We also perform simulations on real-valued data, casting to the complex domain by means of the Hilbert Transform, and verifying the potential interest of CVNN even for non-complex data.
arXiv Detail & Related papers (2023-02-16T13:31:10Z) - A Recursively Recurrent Neural Network (R2N2) Architecture for Learning
Iterative Algorithms [64.3064050603721]
We generalize Runge-Kutta neural network to a recurrent neural network (R2N2) superstructure for the design of customized iterative algorithms.
We demonstrate that regular training of the weight parameters inside the proposed superstructure on input/output data of various computational problem classes yields similar iterations to Krylov solvers for linear equation systems, Newton-Krylov solvers for nonlinear equation systems, and Runge-Kutta solvers for ordinary differential equations.
arXiv Detail & Related papers (2022-11-22T16:30:33Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Spectral Complexity-scaled Generalization Bound of Complex-valued Neural
Networks [78.64167379726163]
This paper is the first work that proves a generalization bound for the complex-valued neural network.
We conduct experiments by training complex-valued convolutional neural networks on different datasets.
arXiv Detail & Related papers (2021-12-07T03:25:25Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - A Survey of Complex-Valued Neural Networks [4.211128681972148]
Artificial neural networks (ANNs) based machine learning models have been widely applied in computer vision, signal processing, wireless communications, and many other domains.
Most of the current implementations of ANNs and machine learning frameworks are using real numbers rather than complex numbers.
There are growing interests in building ANNs using complex numbers, and exploring the potential advantages of the so-called complex-valued neural networks (CVNNs) over their real-valued counterparts.
arXiv Detail & Related papers (2021-01-28T19:40:50Z) - The geometry of integration in text classification RNNs [20.76659136484842]
We study recurrent networks trained on a battery of both natural and synthetic text classification tasks.
We find the dynamics of these trained RNNs to be both interpretable and low-dimensional.
Our observations span multiple architectures and datasets, reflecting a common mechanism RNNs employ to perform text classification.
arXiv Detail & Related papers (2020-10-28T17:58:53Z) - Optimization and Generalization Analysis of Transduction through
Gradient Boosting and Application to Multi-scale Graph Neural Networks [60.22494363676747]
It is known that the current graph neural networks (GNNs) are difficult to make themselves deep due to the problem known as over-smoothing.
Multi-scale GNNs are a promising approach for mitigating the over-smoothing problem.
We derive the optimization and generalization guarantees of transductive learning algorithms that include multi-scale GNNs.
arXiv Detail & Related papers (2020-06-15T17:06:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.