Padding Module: Learning the Padding in Deep Neural Networks
- URL: http://arxiv.org/abs/2301.04608v1
- Date: Wed, 11 Jan 2023 18:03:57 GMT
- Title: Padding Module: Learning the Padding in Deep Neural Networks
- Authors: Fahad Alrasheedi, Xin Zhong, Pei-Chi Huang
- Abstract summary: This paper proposes a trainable Padding Module that can be placed in a deep learning model.
The Padding Module can optimize itself without requiring or influencing the model's entire loss function.
Experiments have shown that the proposed Padding Module outperforms the state-of-the-art competitors and the baseline methods.
- Score: 4.769747792846005
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: During the last decades, many studies have been dedicated to improving the
performance of neural networks, for example, the network architectures,
initialization, and activation. However, investigating the importance and
effects of learnable padding methods in deep learning remains relatively open.
To mitigate the gap, this paper proposes a novel trainable Padding Module that
can be placed in a deep learning model. The Padding Module can optimize itself
without requiring or influencing the model's entire loss function. To train
itself, the Padding Module constructs a ground truth and a predictor from the
inputs by leveraging the underlying structure in the input data for
supervision. As a result, the Padding Module can learn automatically to pad
pixels to the border of its input images or feature maps. The padding contents
are realistic extensions to its input data and simultaneously facilitate the
deep learning model's downstream task. Experiments have shown that the proposed
Padding Module outperforms the state-of-the-art competitors and the baseline
methods. For example, the Padding Module has 1.23% and 0.44% more
classification accuracy than the zero padding when tested on the VGG16 and
ResNet50.
Related papers
- Modular Neural Network Approaches for Surgical Image Recognition [0.0]
We introduce and evaluate different architectures of modular learning for Dorsal Capsulo-Scapholunate Septum (DCSS) instability classification.
Our experiments have shown that modular learning improves performances compared to non-modular systems.
In the second part, we present our approach for data labeling and segmentation with self-training applied on shoulder arthroscopy images.
arXiv Detail & Related papers (2023-07-17T22:28:16Z) - PDSketch: Integrated Planning Domain Programming and Learning [86.07442931141637]
We present a new domain definition language, named PDSketch.
It allows users to flexibly define high-level structures in the transition models.
Details of the transition model will be filled in by trainable neural networks.
arXiv Detail & Related papers (2023-03-09T18:54:12Z) - Neural Attentive Circuits [93.95502541529115]
We introduce a general purpose, yet modular neural architecture called Neural Attentive Circuits (NACs)
NACs learn the parameterization and a sparse connectivity of neural modules without using domain knowledge.
NACs achieve an 8x speedup at inference time while losing less than 3% performance.
arXiv Detail & Related papers (2022-10-14T18:00:07Z) - Emergence of hierarchical modes from deep learning [2.0711789781518752]
We propose a mode decomposition learning that can interpret the weight matrices as a hierarchy of latent modes.
The mode decomposition learning points to a cheap and interpretable route towards the magical deep learning.
arXiv Detail & Related papers (2022-08-21T09:53:32Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Towards Disentangling Information Paths with Coded ResNeXt [11.884259630414515]
We take a novel approach to enhance the transparency of the function of the whole network.
We propose a neural network architecture for classification, in which the information that is relevant to each class flows through specific paths.
arXiv Detail & Related papers (2022-02-10T21:45:49Z) - Neural Network Module Decomposition and Recomposition [35.21448933547118]
We propose a modularization method that decomposes a deep neural network (DNN) into small modules from a functionality perspective.
We demonstrate that the proposed method can decompose and recompose DNNs with high compression ratio and high accuracy.
arXiv Detail & Related papers (2021-12-25T08:36:47Z) - Reasoning-Modulated Representations [85.08205744191078]
We study a common setting where our task is not purely opaque.
Our approach paves the way for a new class of data-efficient representation learning.
arXiv Detail & Related papers (2021-07-19T13:57:13Z) - Neural Function Modules with Sparse Arguments: A Dynamic Approach to
Integrating Information across Layers [84.57980167400513]
Neural Function Modules (NFM) aims to introduce the same structural capability into deep learning.
Most of the work in the context of feed-forward networks combining top-down and bottom-up feedback is limited to classification problems.
The key contribution of our work is to combine attention, sparsity, top-down and bottom-up feedback, in a flexible algorithm.
arXiv Detail & Related papers (2020-10-15T20:43:17Z) - Incremental Training of a Recurrent Neural Network Exploiting a
Multi-Scale Dynamic Memory [79.42778415729475]
We propose a novel incrementally trained recurrent architecture targeting explicitly multi-scale learning.
We show how to extend the architecture of a simple RNN by separating its hidden state into different modules.
We discuss a training algorithm where new modules are iteratively added to the model to learn progressively longer dependencies.
arXiv Detail & Related papers (2020-06-29T08:35:49Z) - Modularizing Deep Learning via Pairwise Learning With Kernels [12.051746916737343]
We present an alternative view on finitely wide, fully trainable deep computation neural networks as stacked linear models in feature spaces.
We then propose a provably optimal modular learning framework for classification that does not require between- module backpropagation.
arXiv Detail & Related papers (2020-05-12T04:19:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.