Shape Adaptor: A Learnable Resizing Module
- URL: http://arxiv.org/abs/2008.00892v2
- Date: Mon, 10 Aug 2020 13:10:50 GMT
- Title: Shape Adaptor: A Learnable Resizing Module
- Authors: Shikun Liu, Zhe Lin, Yilin Wang, Jianming Zhang, Federico Perazzi,
Edward Johns
- Abstract summary: We present a novel resizing module for neural networks: shape adaptor, a drop-in enhancement built on top of traditional resizing layers.
Our implementation enables shape adaptors to be trained end-to-end without any additional supervision.
We show the effectiveness of shape adaptors on two other applications: network compression and transfer learning.
- Score: 59.940372879848624
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We present a novel resizing module for neural networks: shape adaptor, a
drop-in enhancement built on top of traditional resizing layers, such as
pooling, bilinear sampling, and strided convolution. Whilst traditional
resizing layers have fixed and deterministic reshaping factors, our module
allows for a learnable reshaping factor. Our implementation enables shape
adaptors to be trained end-to-end without any additional supervision, through
which network architectures can be optimised for each individual task, in a
fully automated way. We performed experiments across seven image classification
datasets, and results show that by simply using a set of our shape adaptors
instead of the original resizing layers, performance increases consistently
over human-designed networks, across all datasets. Additionally, we show the
effectiveness of shape adaptors on two other applications: network compression
and transfer learning. The source code is available at:
https://github.com/lorenmt/shape-adaptor.
Related papers
- Sample-based Dynamic Hierarchical Transformer with Layer and Head
Flexibility via Contextual Bandit [24.78757412559944]
Transformer requires a fixed number of layers and heads which makes them inflexible to the complexity of individual samples.
We propose a sample-based Dynamic Hierarchical Transformer model whose layers and heads can be dynamically configured with single data samples.
We achieve up to 74% computational savings for both training and inference with a minimal loss of accuracy.
arXiv Detail & Related papers (2023-12-05T15:04:11Z) - Efficient Implementation of a Multi-Layer Gradient-Free Online-Trainable
Spiking Neural Network on FPGA [0.31498833540989407]
ODESA is the first network to have end-to-end multi-layer online local supervised training without using gradients.
This research shows that the network architecture and the online training of weights and thresholds can be implemented efficiently on a large scale in hardware.
arXiv Detail & Related papers (2023-05-31T00:34:15Z) - Augmenting Convolutional networks with attention-based aggregation [55.97184767391253]
We show how to augment any convolutional network with an attention-based global map to achieve non-local reasoning.
We plug this learned aggregation layer with a simplistic patch-based convolutional network parametrized by 2 parameters (width and depth)
It yields surprisingly competitive trade-offs between accuracy and complexity, in particular in terms of memory consumption.
arXiv Detail & Related papers (2021-12-27T14:05:41Z) - Domain Adaptor Networks for Hyperspectral Image Recognition [35.95313368586933]
We consider the problem of adapting a network trained on three-channel color images to a hyperspectral domain with a large number of channels.
We propose domain adaptor networks that map the input to be compatible with a network trained on large-scale color image datasets such as ImageNet.
arXiv Detail & Related papers (2021-08-03T15:06:39Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - Joint Self-Attention and Scale-Aggregation for Self-Calibrated Deraining
Network [13.628218953897946]
In this paper, we propose an effective algorithm, called JDNet, to solve the single image deraining problem.
By designing the Scale-Aggregation and Self-Attention modules with Self-Calibrated convolution skillfully, the proposed model has better deraining results.
arXiv Detail & Related papers (2020-08-06T17:04:34Z) - Learning to Learn Parameterized Classification Networks for Scalable
Input Images [76.44375136492827]
Convolutional Neural Networks (CNNs) do not have a predictable recognition behavior with respect to the input resolution change.
We employ meta learners to generate convolutional weights of main networks for various input scales.
We further utilize knowledge distillation on the fly over model predictions based on different input resolutions.
arXiv Detail & Related papers (2020-07-13T04:27:25Z) - Continual Adaptation for Deep Stereo [52.181067640300014]
We propose a continual adaptation paradigm for deep stereo networks designed to deal with challenging and ever-changing environments.
In our paradigm, the learning signals needed to continuously adapt models online can be sourced from self-supervision via right-to-left image warping or from traditional stereo algorithms.
Our network architecture and adaptation algorithms realize the first real-time self-adaptive deep stereo system.
arXiv Detail & Related papers (2020-07-10T08:15:58Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.