Squeeze aggregated excitation network
- URL: http://arxiv.org/abs/2308.13343v1
- Date: Fri, 25 Aug 2023 12:30:48 GMT
- Title: Squeeze aggregated excitation network
- Authors: Mahendran N
- Abstract summary: Convolutional neural networks have spatial representations which read patterns in the vision tasks.
We propose SaEnet, Squeeze aggregated excitation network, for learning global channelwise representation in between layers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Convolutional neural networks have spatial representations which read
patterns in the vision tasks. Squeeze and excitation links the channel wise
representations by explicitly modeling on channel level. Multi layer
perceptrons learn global representations and in most of the models it is used
often at the end after all convolutional layers to gather all the information
learned before classification. We propose a method of inducing the global
representations within channels to have better performance of the model. We
propose SaEnet, Squeeze aggregated excitation network, for learning global
channelwise representation in between layers. The proposed module takes
advantage of passing important information after squeeze by having aggregated
excitation before regaining its shape. We also introduce a new idea of having a
multibranch linear(dense) layer in the network. This learns global
representations from the condensed information which enhances the
representational power of the network. The proposed module have undergone
extensive experiments by using Imagenet and CIFAR100 datasets and compared with
closely related architectures. The analyzes results that proposed models
outputs are comparable and in some cases better than existing state of the art
architectures.
Related papers
- Linking in Style: Understanding learned features in deep learning models [0.0]
Convolutional neural networks (CNNs) learn abstract features to perform object classification.
We propose an automatic method to visualize and systematically analyze learned features in CNNs.
arXiv Detail & Related papers (2024-09-25T12:28:48Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Neural Clustering based Visual Representation Learning [61.72646814537163]
Clustering is one of the most classic approaches in machine learning and data analysis.
We propose feature extraction with clustering (FEC), which views feature extraction as a process of selecting representatives from data.
FEC alternates between grouping pixels into individual clusters to abstract representatives and updating the deep features of pixels with current representatives.
arXiv Detail & Related papers (2024-03-26T06:04:50Z) - SENetV2: Aggregated dense layer for channelwise and global
representations [0.0]
We introduce a novel aggregated multilayer perceptron, a multi-branch dense layer, within the Squeeze residual module.
This fusion enhances the network's ability to capture channel-wise patterns and have global knowledge.
We conduct extensive experiments on benchmark datasets to validate the model and compare them with established architectures.
arXiv Detail & Related papers (2023-11-17T14:10:57Z) - Edge-Enhanced Global Disentangled Graph Neural Network for Sequential
Recommendation [44.15486708923762]
We propose an Edge-Enhanced Global Disentangled Graph Neural Network (EGD-GNN) model to capture the relation information between items.
At the global level, we build a global-link graph over all sequences to model item relationships.
At the local level, we apply a variational auto-encoder framework to learn user intention over the current sequence.
arXiv Detail & Related papers (2021-11-20T08:15:20Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Do Vision Transformers See Like Convolutional Neural Networks? [45.69780772718875]
Recent work has shown that (Vision) Transformer models (ViT) can achieve comparable or even superior performance on image classification tasks.
Are they acting like convolutional networks, or learning entirely different visual representations?
We find striking differences between the two architectures, such as ViT having more uniform representations across all layers.
arXiv Detail & Related papers (2021-08-19T17:27:03Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - Incremental Training of a Recurrent Neural Network Exploiting a
Multi-Scale Dynamic Memory [79.42778415729475]
We propose a novel incrementally trained recurrent architecture targeting explicitly multi-scale learning.
We show how to extend the architecture of a simple RNN by separating its hidden state into different modules.
We discuss a training algorithm where new modules are iteratively added to the model to learn progressively longer dependencies.
arXiv Detail & Related papers (2020-06-29T08:35:49Z) - Embedded Encoder-Decoder in Convolutional Networks Towards Explainable
AI [0.0]
This paper proposes a new explainable convolutional neural network (XCNN) which represents important and driving visual features of stimuli.
The experimental results on the CIFAR-10, Tiny ImageNet, and MNIST datasets showed the success of our algorithm (XCNN) to make CNNs explainable.
arXiv Detail & Related papers (2020-06-19T15:49:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.