Lightweight Convolutional Neural Networks By Hypercomplex
Parameterization
- URL: http://arxiv.org/abs/2110.04176v1
- Date: Fri, 8 Oct 2021 14:57:19 GMT
- Title: Lightweight Convolutional Neural Networks By Hypercomplex
Parameterization
- Authors: Eleonora Grassucci, Aston Zhang, Danilo Comminiello
- Abstract summary: We define the parameterization of hypercomplex convolutional layers to develop lightweight and efficient large-scale convolutional models.
Our method grasps the convolution rules and the filters organization directly from data.
We demonstrate the versatility of this approach to multiple domains of application by performing experiments on various image datasets and audio datasets.
- Score: 10.420215908252425
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hypercomplex neural networks have proved to reduce the overall number of
parameters while ensuring valuable performances by leveraging the properties of
Clifford algebras. Recently, hypercomplex linear layers have been further
improved by involving efficient parameterized Kronecker products. In this
paper, we define the parameterization of hypercomplex convolutional layers to
develop lightweight and efficient large-scale convolutional models. Our method
grasps the convolution rules and the filters organization directly from data
without requiring a rigidly predefined domain structure to follow. The proposed
approach is flexible to operate in any user-defined or tuned domain, from 1D to
$n$D regardless of whether the algebra rules are preset. Such a malleability
allows processing multidimensional inputs in their natural domain without
annexing further dimensions, as done, instead, in quaternion neural networks
for 3D inputs like color images. As a result, the proposed method operates with
$1/n$ free parameters as regards its analog in the real domain. We demonstrate
the versatility of this approach to multiple domains of application by
performing experiments on various image datasets as well as audio datasets in
which our method outperforms real and quaternion-valued counterparts.
Related papers
- Multi-Grid Tensorized Fourier Neural Operator for High-Resolution PDEs [93.82811501035569]
We introduce a new data efficient and highly parallelizable operator learning approach with reduced memory requirement and better generalization.
MG-TFNO scales to large resolutions by leveraging local and global structures of full-scale, real-world phenomena.
We demonstrate superior performance on the turbulent Navier-Stokes equations where we achieve less than half the error with over 150x compression.
arXiv Detail & Related papers (2023-09-29T20:18:52Z) - Tunable Convolutions with Parametric Multi-Loss Optimization [5.658123802733283]
Behavior of neural networks is irremediably determined by the specific loss and data used during training.
It is often desirable to tune the model at inference time based on external factors such as preferences of the user or dynamic characteristics of the data.
This is especially important to balance the perception-distortion trade-off of ill-posed image-to-image translation tasks.
arXiv Detail & Related papers (2023-04-03T11:36:10Z) - Efficient Parametric Approximations of Neural Network Function Space
Distance [6.117371161379209]
It is often useful to compactly summarize important properties of model parameters and training data so that they can be used later without storing and/or iterating over the entire dataset.
We consider estimating the Function Space Distance (FSD) over a training set, i.e. the average discrepancy between the outputs of two neural networks.
We propose a Linearized Activation TRick (LAFTR) and derive an efficient approximation to FSD for ReLU neural networks.
arXiv Detail & Related papers (2023-02-07T15:09:23Z) - Deep Diversity-Enhanced Feature Representation of Hyperspectral Images [87.47202258194719]
We rectify 3D convolution by modifying its topology to enhance the rank upper-bound.
We also propose a novel diversity-aware regularization (DA-Reg) term that acts on the feature maps to maximize independence among elements.
To demonstrate the superiority of the proposed Re$3$-ConvSet and DA-Reg, we apply them to various HS image processing and analysis tasks.
arXiv Detail & Related papers (2023-01-15T16:19:18Z) - Hyperparameter Tuning in Echo State Networks [0.0]
Echo State Networks are a type of recurrent neural network with a large randomly generated reservoir and a small number of readout connections trained via linear regression.
We propose an alternative approach of hyper parameter tuning based on the Covariance Matrix Adaptation Evolution Strategy (CMA-ES)
arXiv Detail & Related papers (2022-07-16T16:20:01Z) - CSformer: Bridging Convolution and Transformer for Compressive Sensing [65.22377493627687]
This paper proposes a hybrid framework that integrates the advantages of leveraging detailed spatial information from CNN and the global context provided by transformer for enhanced representation learning.
The proposed approach is an end-to-end compressive image sensing method, composed of adaptive sampling and recovery.
The experimental results demonstrate the effectiveness of the dedicated transformer-based architecture for compressive sensing.
arXiv Detail & Related papers (2021-12-31T04:37:11Z) - Dynamic Convolution for 3D Point Cloud Instance Segmentation [146.7971476424351]
We propose an approach to instance segmentation from 3D point clouds based on dynamic convolution.
We gather homogeneous points that have identical semantic categories and close votes for the geometric centroids.
The proposed approach is proposal-free, and instead exploits a convolution process that adapts to the spatial and semantic characteristics of each instance.
arXiv Detail & Related papers (2021-07-18T09:05:16Z) - Beyond Fully-Connected Layers with Quaternions: Parameterization of
Hypercomplex Multiplications with $1/n$ Parameters [71.09633069060342]
We propose parameterizing hypercomplex multiplications, allowing models to learn multiplication rules from data regardless of whether such rules are predefined.
Our method not only subsumes the Hamilton product, but also learns to operate on any arbitrary nD hypercomplex space.
arXiv Detail & Related papers (2021-02-17T06:16:58Z) - DyCo3D: Robust Instance Segmentation of 3D Point Clouds through Dynamic
Convolution [136.7261709896713]
We propose a data-driven approach that generates the appropriate convolution kernels to apply in response to the nature of the instances.
The proposed method achieves promising results on both ScanetNetV2 and S3DIS.
It also improves inference speed by more than 25% over the current state-of-the-art.
arXiv Detail & Related papers (2020-11-26T14:56:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.