1-Dimensional polynomial neural networks for audio signal related
problems
- URL: http://arxiv.org/abs/2009.04077v2
- Date: Wed, 12 Jan 2022 19:07:17 GMT
- Title: 1-Dimensional polynomial neural networks for audio signal related
problems
- Authors: Habib Ben Abdallah, Christopher J. Henry, Sheela Ramanna
- Abstract summary: We show that the proposed model can extract more relevant information from the data than a 1DCNN in less time and with less memory.
We show that this non-linearity enables the model to yield better results with less computational and spatial complexity than a regular 1DCNN on various classification and regression problems related to audio signals.
- Score: 3.867363075280544
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In addition to being extremely non-linear, modern problems require millions
if not billions of parameters to solve or at least to get a good approximation
of the solution, and neural networks are known to assimilate that complexity by
deepening and widening their topology in order to increase the level of
non-linearity needed for a better approximation. However, compact topologies
are always preferred to deeper ones as they offer the advantage of using less
computational units and less parameters. This compacity comes at the price of
reduced non-linearity and thus, of limited solution search space. We propose
the 1-Dimensional Polynomial Neural Network (1DPNN) model that uses automatic
polynomial kernel estimation for 1-Dimensional Convolutional Neural Networks
(1DCNNs) and that introduces a high degree of non-linearity from the first
layer which can compensate the need for deep and/or wide topologies. We show
that this non-linearity enables the model to yield better results with less
computational and spatial complexity than a regular 1DCNN on various
classification and regression problems related to audio signals, even though it
introduces more computational and spatial complexity on a neuronal level. The
experiments were conducted on three publicly available datasets and demonstrate
that, on the problems that were tackled, the proposed model can extract more
relevant information from the data than a 1DCNN in less time and with less
memory.
Related papers
- Sequencing the Neurome: Towards Scalable Exact Parameter Reconstruction of Black-Box Neural Networks [7.0710630443004705]
Inferring exact parameters of a neural network with only query access is an NP-Hard problem.
We present a novel query generation algorithm that produces maximally informative samples, letting us untangle the non-linear relationships efficiently.
We demonstrate reconstruction of a hidden network containing over 1.5 million parameters, and of one 7 layers deep, the largest and deepest reconstructions to date, with max parameter difference less than 0.0001.
arXiv Detail & Related papers (2024-09-27T21:02:04Z) - Stacked tensorial neural networks for reduced-order modeling of a
parametric partial differential equation [0.0]
I describe a deep neural network architecture that fuses multiple TNNs into a larger network.
I evaluate this architecture on a parametric PDE with three independent variables and three parameters.
arXiv Detail & Related papers (2023-12-21T21:44:50Z) - DEMAND: Deep Matrix Approximately NonlinearDecomposition to Identify
Meta, Canonical, and Sub-Spatial Pattern of functional Magnetic Resonance
Imaging in the Human Brain [8.93274096260726]
We propose a novel deep nonlinear matrix factorization named Deep Approximately Decomposition (DEMAND) in this work to take advantage of the shallow linear model, e.g., Sparse Dictionary Learning (SDL) and Deep Neural Networks (DNNs)
DEMAND can reveal the reproducible meta, canonical, and sub-spatial features of the human brain more efficiently than other peer methodologies.
arXiv Detail & Related papers (2022-05-20T15:55:01Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Measuring Model Complexity of Neural Networks with Curve Activation
Functions [100.98319505253797]
We propose the linear approximation neural network (LANN) to approximate a given deep model with curve activation function.
We experimentally explore the training process of neural networks and detect overfitting.
We find that the $L1$ and $L2$ regularizations suppress the increase of model complexity.
arXiv Detail & Related papers (2020-06-16T07:38:06Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.