A General Framework for Hypercomplex-valued Extreme Learning Machines
- URL: http://arxiv.org/abs/2101.06166v1
- Date: Fri, 15 Jan 2021 15:22:05 GMT
- Title: A General Framework for Hypercomplex-valued Extreme Learning Machines
- Authors: Guilherme Vieira and Marcos Eduardo Valle
- Abstract summary: This paper aims to establish a framework for extreme learning machines (ELMs) on general hypercomplex algebras.
We show a framework to operate in these algebras through real-valued linear algebra operations.
Experiments highlight the excellent performance of hypercomplex-valued ELMs to treat high-dimensional data.
- Score: 2.055949720959582
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper aims to establish a framework for extreme learning machines (ELMs)
on general hypercomplex algebras. Hypercomplex neural networks are machine
learning models that feature higher-dimension numbers as parameters, inputs,
and outputs. Firstly, we review broad hypercomplex algebras and show a
framework to operate in these algebras through real-valued linear algebra
operations in a robust manner. We proceed to explore a handful of well-known
four-dimensional examples. Then, we propose the hypercomplex-valued ELMs and
derive their learning using a hypercomplex-valued least-squares problem.
Finally, we compare real and hypercomplex-valued ELM models' performance in an
experiment on time-series prediction and another on color image auto-encoding.
The computational experiments highlight the excellent performance of
hypercomplex-valued ELMs to treat high-dimensional data, including models based
on unusual hypercomplex algebras.
Related papers
- Learning Linear Attention in Polynomial Time [115.68795790532289]
We provide the first results on learnability of single-layer Transformers with linear attention.
We show that linear attention may be viewed as a linear predictor in a suitably defined RKHS.
We show how to efficiently identify training datasets for which every empirical riskr is equivalent to the linear Transformer.
arXiv Detail & Related papers (2024-10-14T02:41:01Z) - Revisiting time-variant complex conjugate matrix equations with their corresponding real field time-variant large-scale linear equations, neural hypercomplex numbers space compressive approximation approach [1.1970409518725493]
Time-variant complex conjugate matrix equations need to be transformed into corresponding real field time-variant large-scale linear equations.
In this paper, zeroing neural dynamic models based on complex field error (called Con-CZND1) and based on real field error (called Con-CZND2) are proposed.
Numerical experiments verify Con-CZND1 conj model effectiveness and highlight NHNSCAA importance.
arXiv Detail & Related papers (2024-08-26T07:33:45Z) - Demystifying the Hypercomplex: Inductive Biases in Hypercomplex Deep Learning [23.501824517684465]
This paper provides a framework for understanding why hypercomplex deep learning methods are so successful and how their potential can be exploited.
We show that it is possible to derive specific inductive biases in the hypercomplex domains.
These biases prove effective in managing the distinctive properties of these domains, as well as the complex structures of multidimensional and multimodal signals.
arXiv Detail & Related papers (2024-05-11T14:41:48Z) - CoLA: Exploiting Compositional Structure for Automatic and Efficient
Numerical Linear Algebra [62.37017125812101]
We propose a simple but general framework for large-scale linear algebra problems in machine learning, named CoLA.
By combining a linear operator abstraction with compositional dispatch rules, CoLA automatically constructs memory and runtime efficient numerical algorithms.
We showcase its efficacy across a broad range of applications, including partial differential equations, Gaussian processes, equivariant model construction, and unsupervised learning.
arXiv Detail & Related papers (2023-09-06T14:59:38Z) - Sample Complexity of Kernel-Based Q-Learning [11.32718794195643]
We propose a non-parametric Q-learning algorithm which finds an $epsilon$-optimal policy in an arbitrarily large scale discounted MDP.
To the best of our knowledge, this is the first result showing a finite sample complexity under such a general model.
arXiv Detail & Related papers (2023-02-01T19:46:25Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - An Extension to Basis-Hypervectors for Learning from Circular Data in
Hyperdimensional Computing [62.997667081978825]
Hyperdimensional Computing (HDC) is a computation framework based on properties of high-dimensional random spaces.
We present a study on basis-hypervector sets, which leads to practical contributions to HDC in general.
We introduce a method to learn from circular data, an important type of information never before addressed in machine learning with HDC.
arXiv Detail & Related papers (2022-05-16T18:04:55Z) - HyperNP: Interactive Visual Exploration of Multidimensional Projection
Hyperparameters [61.354362652006834]
HyperNP is a scalable method that allows for real-time interactive exploration of projection methods by training neural network approximations.
We evaluate the performance of the HyperNP across three datasets in terms of performance and speed.
arXiv Detail & Related papers (2021-06-25T17:28:14Z) - Parameterized Hypercomplex Graph Neural Networks for Graph
Classification [1.1852406625172216]
We develop graph neural networks that leverage the properties of hypercomplex feature transformation.
In particular, in our proposed class of models, the multiplication rule specifying the algebra itself is inferred from the data during training.
We test our proposed hypercomplex GNN on several open graph benchmark datasets and show that our models reach state-of-the-art performance.
arXiv Detail & Related papers (2021-03-30T18:01:06Z) - Beyond Fully-Connected Layers with Quaternions: Parameterization of
Hypercomplex Multiplications with $1/n$ Parameters [71.09633069060342]
We propose parameterizing hypercomplex multiplications, allowing models to learn multiplication rules from data regardless of whether such rules are predefined.
Our method not only subsumes the Hamilton product, but also learns to operate on any arbitrary nD hypercomplex space.
arXiv Detail & Related papers (2021-02-17T06:16:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.