An Interpretive Constrained Linear Model for ResNet and MgNet
- URL: http://arxiv.org/abs/2112.07441v1
- Date: Tue, 14 Dec 2021 14:52:44 GMT
- Title: An Interpretive Constrained Linear Model for ResNet and MgNet
- Authors: Juncai He, Jinchao Xu, Lian Zhang, Jianqing Zhu
- Abstract summary: We propose a constrained linear data-feature-mapping model as an interpretable mathematical model for image classification using a convolutional neural network (CNN)
We establish detailed connections between the traditional iterative schemes for linear systems and the architectures of the basic blocks of ResNet- and MgNet-type models.
We present some modified ResNet models that compared with the original models have fewer parameters and yet can produce more accurate results.
- Score: 4.407784399315197
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a constrained linear data-feature-mapping model as an
interpretable mathematical model for image classification using a convolutional
neural network (CNN). From this viewpoint, we establish detailed connections
between the traditional iterative schemes for linear systems and the
architectures of the basic blocks of ResNet- and MgNet-type models. Using these
connections, we present some modified ResNet models that compared with the
original models have fewer parameters and yet can produce more accurate
results, thereby demonstrating the validity of this constrained learning
data-feature-mapping assumption. Based on this assumption, we further propose a
general data-feature iterative scheme to show the rationality of MgNet. We also
provide a systematic numerical study on MgNet to show its success and
advantages in image classification problems and demonstrate its advantages in
comparison with established networks.
Related papers
- Generalized Factor Neural Network Model for High-dimensional Regression [50.554377879576066]
We tackle the challenges of modeling high-dimensional data sets with latent low-dimensional structures hidden within complex, non-linear, and noisy relationships.
Our approach enables a seamless integration of concepts from non-parametric regression, factor models, and neural networks for high-dimensional regression.
arXiv Detail & Related papers (2025-02-16T23:13:55Z) - Telematics Combined Actuarial Neural Networks for Cross-Sectional and
Longitudinal Claim Count Data [0.0]
We present novel cross-sectional and longitudinal claim count models for vehicle insurance built upon the Combined Actuarial Network (CANN) framework proposed by Mario W"uthrich and Michael Merz.
arXiv Detail & Related papers (2023-08-03T12:40:44Z) - MAgNET: A Graph U-Net Architecture for Mesh-Based Simulations [0.5185522256407782]
We present MAgNET, which extends the well-known convolutional neural networks to accommodate arbitrary graph-structured data.
We demonstrate the predictive capabilities of MAgNET in surrogate modeling for non-linear finite element simulations in the mechanics of solids.
arXiv Detail & Related papers (2022-11-01T19:23:45Z) - Graph Polynomial Convolution Models for Node Classification of
Non-Homophilous Graphs [52.52570805621925]
We investigate efficient learning from higher-order graph convolution and learning directly from adjacency matrix for node classification.
We show that the resulting model lead to new graphs and residual scaling parameter.
We demonstrate that the proposed methods obtain improved accuracy for node-classification of non-homophilous parameters.
arXiv Detail & Related papers (2022-09-12T04:46:55Z) - Recurrent Graph Tensor Networks: A Low-Complexity Framework for
Modelling High-Dimensional Multi-Way Sequence [24.594587557319837]
We develop a graph filter framework for approximating the modelling of hidden states in Recurrent Neural Networks (RNNs)
The proposed framework is validated through several multi-way sequence modelling tasks and benchmarked against traditional RNNs.
We show that the proposed RGTN is capable of not only out-performing standard RNNs, but also mitigating the Curse of Dimensionality associated with traditional RNNs.
arXiv Detail & Related papers (2020-09-18T10:13:36Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - Kernel and Rich Regimes in Overparametrized Models [69.40899443842443]
We show that gradient descent on overparametrized multilayer networks can induce rich implicit biases that are not RKHS norms.
We also demonstrate this transition empirically for more complex matrix factorization models and multilayer non-linear networks.
arXiv Detail & Related papers (2020-02-20T15:43:02Z) - Revisiting Graph based Collaborative Filtering: A Linear Residual Graph
Convolutional Network Approach [55.44107800525776]
Graph Convolutional Networks (GCNs) are state-of-the-art graph based representation learning models.
In this paper, we revisit GCN based Collaborative Filtering (CF) based Recommender Systems (RS)
We show that removing non-linearities would enhance recommendation performance, consistent with the theories in simple graph convolutional networks.
We propose a residual network structure that is specifically designed for CF with user-item interaction modeling.
arXiv Detail & Related papers (2020-01-28T04:41:25Z) - On generalized residue network for deep learning of unknown dynamical
systems [3.350695583277162]
We present a general numerical approach for learning unknown dynamical systems using deep neural networks (DNNs)
Our method is built upon recent studies that identified the residue network (ResNet) as an effective neural network structure.
arXiv Detail & Related papers (2020-01-23T01:50:22Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.