CVKAN: Complex-Valued Kolmogorov-Arnold Networks
- URL: http://arxiv.org/abs/2502.02417v3
- Date: Tue, 22 Apr 2025 15:09:46 GMT
- Title: CVKAN: Complex-Valued Kolmogorov-Arnold Networks
- Authors: Matthias Wolff, Florian Eilers, Xiaoyi Jiang,
- Abstract summary: We show how to transfer a complex-valued Kolmogorov-Arnold Network into the complex domain.<n>Our proposed CVKAN is more stable and performs on par or better than real-valued KANs.
- Score: 6.095572174539791
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work we propose CVKAN, a complex-valued Kolmogorov-Arnold Network (KAN), to join the intrinsic interpretability of KANs and the advantages of Complex-Valued Neural Networks (CVNNs). We show how to transfer a KAN and the necessary associated mechanisms into the complex domain. To confirm that CVKAN meets expectations we conduct experiments on symbolic complex-valued function fitting and physically meaningful formulae as well as on a more realistic dataset from knot theory. Our proposed CVKAN is more stable and performs on par or better than real-valued KANs while requiring less parameters and a shallower network architecture, making it more explainable.
Related papers
- Hybrid Real- and Complex-valued Neural Network Architecture [2.6739705603496327]
We propose a emphhybrid real- and complex-valued emphneural network (HNN) architecture, designed to combine the computational efficiency of real-valued processing with the ability to handle complex-valued data.
Experiments with the AudioMNIST dataset demonstrate that the HNN reduces cross-entropy loss and consumes less parameters compared to an RVNN for all considered cases.
arXiv Detail & Related papers (2025-04-04T14:52:44Z) - PRKAN: Parameter-Reduced Kolmogorov-Arnold Networks [47.947045173329315]
Kolmogorov-Arnold Networks (KANs) represent an innovation in neural network architectures.<n>KANs offer a compelling alternative to Multi-Layer Perceptrons (MLPs) in models such as CNNs, RecurrentReduced Networks (RNNs) and Transformers.<n>This paper introduces PRKANs, which employ several methods to reduce the parameter count in layers, making them comparable to Neural M layers.
arXiv Detail & Related papers (2025-01-13T03:07:39Z) - KAE: Kolmogorov-Arnold Auto-Encoder for Representation Learning [2.6713407440802253]
The Kolmogorov-Arnold Network (KAN) has recently gained attention as an alternative to traditional multi-layer perceptrons (MLPs)<n>In this paper, we introduce the Kolmogorov-Arnold AutoEncoder (KAE), which integrates KAN with autoencoders (AEs)<n> Experiments on benchmark datasets demonstrate that KAE improves latent representation quality, reduces reconstruction errors, and achieves superior performance in downstream tasks.
arXiv Detail & Related papers (2024-12-31T12:49:03Z) - KANICE: Kolmogorov-Arnold Networks with Interactive Convolutional Elements [1.663204995903499]
We introduce KANICE, a novel neural architecture that combines Convolutional Neural Networks (CNNs) with Kolmogorov-Arnold Network (KAN) principles.
KANICE integrates Interactive Convolutional Blocks (ICBs) and KAN linear layers into a CNN framework.
We evaluated KANICE on four datasets: MNIST, Fashion-MNIST, EMNIST, and SVHN.
arXiv Detail & Related papers (2024-10-22T16:50:34Z) - Kolmogorov-Arnold Network Autoencoders [0.0]
Kolmogorov-Arnold Networks (KANs) are promising alternatives to Multi-Layer Perceptrons (MLPs)
KANs align closely with the Kolmogorov-Arnold representation theorem, potentially enhancing both model accuracy and interpretability.
Our results demonstrate that KAN-based autoencoders achieve competitive performance in terms of reconstruction accuracy.
arXiv Detail & Related papers (2024-10-02T22:56:00Z) - Kolmogorov-Smirnov GAN [52.36633001046723]
We propose a novel deep generative model, the Kolmogorov-Smirnov Generative Adversarial Network (KSGAN)
Unlike existing approaches, KSGAN formulates the learning process as a minimization of the Kolmogorov-Smirnov (KS) distance.
arXiv Detail & Related papers (2024-06-28T14:30:14Z) - Convolutional Kolmogorov-Arnold Networks [41.94295877935867]
We present Convolutional Kolmogorov-Arnold Networks (KANs)
KANs replace traditional fixed-weight kernels with learnable non-linear functions.
We empirically evaluate Convolutional KANs on the Fashion-MNIST dataset, demonstrating competitive accuracy with up to 50% fewer parameters compared to baseline CNNs.
arXiv Detail & Related papers (2024-06-19T02:09:44Z) - Smooth Kolmogorov Arnold networks enabling structural knowledge representation [0.0]
Kolmogorov-Arnold Networks (KANs) offer an efficient and interpretable alternative to traditional multi-layer perceptron (MLP) architectures.
By leveraging inherent structural knowledge, KANs may reduce the data required for training and mitigate the risk of generating hallucinated predictions.
arXiv Detail & Related papers (2024-05-18T15:27:14Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Towards Understanding Theoretical Advantages of Complex-Reaction
Networks [77.34726150561087]
We show that a class of functions can be approximated by a complex-reaction network using the number of parameters.
For empirical risk minimization, our theoretical result shows that the critical point set of complex-reaction networks is a proper subset of that of real-valued networks.
arXiv Detail & Related papers (2021-08-15T10:13:49Z) - Over-and-Under Complete Convolutional RNN for MRI Reconstruction [57.95363471940937]
Recent deep learning-based methods for MR image reconstruction usually leverage a generic auto-encoder architecture.
We propose an Over-and-Under Complete Convolu?tional Recurrent Neural Network (OUCR), which consists of an overcomplete and an undercomplete Convolutional Recurrent Neural Network(CRNN)
The proposed method achieves significant improvements over the compressed sensing and popular deep learning-based methods with less number of trainable parameters.
arXiv Detail & Related papers (2021-06-16T15:56:34Z) - Neural Networks Enhancement with Logical Knowledge [83.9217787335878]
We propose an extension of KENN for relational data.
The results show that KENN is capable of increasing the performances of the underlying neural network even in the presence relational data.
arXiv Detail & Related papers (2020-09-13T21:12:20Z) - Shifu2: A Network Representation Learning Based Model for
Advisor-advisee Relationship Mining [82.75996880087747]
We propose a novel model based on Network Representation Learning (NRL), namely Shifu2.
Shifu2 takes the collaboration network as input and the identified advisor-advisee relationship as output.
We generate a large-scale academic genealogy dataset by taking advantage of Shifu2.
arXiv Detail & Related papers (2020-08-17T05:40:06Z) - Toward fast and accurate human pose estimation via soft-gated skip
connections [97.06882200076096]
This paper is on highly accurate and highly efficient human pose estimation.
We re-analyze this design choice in the context of improving both the accuracy and the efficiency over the state-of-the-art.
Our model achieves state-of-the-art results on the MPII and LSP datasets.
arXiv Detail & Related papers (2020-02-25T18:51:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.