Piecewise Linear Neural Networks and Deep Learning
- URL: http://arxiv.org/abs/2206.09149v1
- Date: Sat, 18 Jun 2022 08:41:42 GMT
- Title: Piecewise Linear Neural Networks and Deep Learning
- Authors: Qinghua Tao, Li Li, Xiaolin Huang, Xiangming Xi, Shuning Wang, Johan
A.K. Suykens
- Abstract summary: PieceWise Linear Neural Networks (PWLNNs) have proven successful in various fields, most recently in deep learning.
In 1977, the canonical representation pioneered the works of shallow PWLNNs learned by incremental designs.
In 2010, the Rectified Linear Unit (ReLU) advocated the prevalence of PWLNNs in deep learning.
- Score: 27.02556725989978
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: As a powerful modelling method, PieceWise Linear Neural Networks (PWLNNs)
have proven successful in various fields, most recently in deep learning. To
apply PWLNN methods, both the representation and the learning have long been
studied. In 1977, the canonical representation pioneered the works of shallow
PWLNNs learned by incremental designs, but the applications to large-scale data
were prohibited. In 2010, the Rectified Linear Unit (ReLU) advocated the
prevalence of PWLNNs in deep learning. Ever since, PWLNNs have been
successfully applied to extensive tasks and achieved advantageous performances.
In this Primer, we systematically introduce the methodology of PWLNNs by
grouping the works into shallow and deep networks. Firstly, different PWLNN
representation models are constructed with elaborated examples. With PWLNNs,
the evolution of learning algorithms for data is presented and fundamental
theoretical analysis follows up for in-depth understandings. Then,
representative applications are introduced together with discussions and
outlooks.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Multiway Multislice PHATE: Visualizing Hidden Dynamics of RNNs through Training [6.326396282553267]
Recurrent neural networks (RNNs) are a widely used tool for sequential data analysis, however, they are still often seen as black boxes of computation.
Here, we present Multiway Multislice PHATE (MM-PHATE), a novel method for visualizing the evolution of RNNs' hidden states.
arXiv Detail & Related papers (2024-06-04T05:05:27Z) - Topological Representations of Heterogeneous Learning Dynamics of Recurrent Spiking Neural Networks [16.60622265961373]
Spiking Neural Networks (SNNs) have become an essential paradigm in neuroscience and artificial intelligence.
Recent advances in literature have studied the network representations of deep neural networks.
arXiv Detail & Related papers (2024-03-19T05:37:26Z) - When Deep Learning Meets Polyhedral Theory: A Survey [6.899761345257773]
In the past decade, deep became the prevalent methodology for predictive modeling thanks to the remarkable accuracy of deep neural learning.
Meanwhile, the structure of neural networks converged back to simplerwise and linear functions.
arXiv Detail & Related papers (2023-04-29T11:46:53Z) - What can linearized neural networks actually say about generalization? [67.83999394554621]
In certain infinitely-wide neural networks, the neural tangent kernel (NTK) theory fully characterizes generalization.
We show that the linear approximations can indeed rank the learning complexity of certain tasks for neural networks.
Our work provides concrete examples of novel deep learning phenomena which can inspire future theoretical research.
arXiv Detail & Related papers (2021-06-12T13:05:11Z) - A self consistent theory of Gaussian Processes captures feature learning
effects in finite CNNs [2.28438857884398]
Deep neural networks (DNNs) in the infinite width/channel limit have received much attention recently.
Despite their theoretical appeal, this viewpoint lacks a crucial ingredient of deep learning in finite DNNs, laying at the heart of their success -- feature learning.
Here we consider DNNs trained with noisy gradient descent on a large training set and derive a self consistent Gaussian Process theory accounting for strong finite-DNN and feature learning effects.
arXiv Detail & Related papers (2021-06-08T05:20:00Z) - Recent Advances in Large Margin Learning [63.982279380483526]
This paper serves as a survey of recent advances in large margin training and its theoretical foundations, mostly for (nonlinear) deep neural networks (DNNs)
We generalize the formulation of classification margins from classical research to latest DNNs, summarize theoretical connections between the margin, network generalization, and robustness, and introduce recent efforts in enlarging the margins for DNNs comprehensively.
arXiv Detail & Related papers (2021-03-25T04:12:00Z) - A Practical Tutorial on Graph Neural Networks [49.919443059032226]
Graph neural networks (GNNs) have recently grown in popularity in the field of artificial intelligence (AI)
This tutorial exposes the power and novelty of GNNs to AI practitioners.
arXiv Detail & Related papers (2020-10-11T12:36:17Z) - How Neural Networks Extrapolate: From Feedforward to Graph Neural
Networks [80.55378250013496]
We study how neural networks trained by gradient descent extrapolate what they learn outside the support of the training distribution.
Graph Neural Networks (GNNs) have shown some success in more complex tasks.
arXiv Detail & Related papers (2020-09-24T17:48:59Z) - Neural Networks Enhancement with Logical Knowledge [83.9217787335878]
We propose an extension of KENN for relational data.
The results show that KENN is capable of increasing the performances of the underlying neural network even in the presence relational data.
arXiv Detail & Related papers (2020-09-13T21:12:20Z) - Deep Transform and Metric Learning Network: Wedding Deep Dictionary
Learning and Neural Networks [34.49034775978504]
We propose a novel DDL approach where each DL layer can be formulated as a combination of one linear layer and a Recurrent Neural Network (RNN)
Our proposed work unveils new insights into Neural Networks and DDL and provides a new, efficient and competitive approach to jointly learn a deep transform and a metric for inference applications.
arXiv Detail & Related papers (2020-02-18T22:04:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.