Multipod Convolutional Network
- URL: http://arxiv.org/abs/2210.00689v1
- Date: Mon, 3 Oct 2022 02:37:57 GMT
- Title: Multipod Convolutional Network
- Authors: Hongyi Pan, Salih Atici, Ahmet Enis Cetin
- Abstract summary: We experimentally observed that three parallel pod networks (TripodNet) produce the best results in commonly used object recognition datasets.
TripodNet achieved state-of-the-art performance on CIFAR-10 and ImageNet datasets.
- Score: 2.1485350418225244
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we introduce a convolutional network which we call MultiPodNet
consisting of a combination of two or more convolutional networks which process
the input image in parallel to achieve the same goal. Output feature maps of
parallel convolutional networks are fused at the fully connected layer of the
network. We experimentally observed that three parallel pod networks
(TripodNet) produce the best results in commonly used object recognition
datasets. Baseline pod networks can be of any type. In this paper, we use
ResNets as baseline networks and their inputs are augmented image patches. The
number of parameters of the TripodNet is about three times that of a single
ResNet. We train the TripodNet using the standard backpropagation type
algorithms. In each individual ResNet, parameters are initialized with
different random numbers during training. The TripodNet achieved
state-of-the-art performance on CIFAR-10 and ImageNet datasets. For example, it
improved the accuracy of a single ResNet from 91.66% to 92.47% under the same
training process on the CIFAR-10 dataset.
Related papers
- SVNet: Where SO(3) Equivariance Meets Binarization on Point Cloud
Representation [65.4396959244269]
The paper tackles the challenge by designing a general framework to construct 3D learning architectures.
The proposed approach can be applied to general backbones like PointNet and DGCNN.
Experiments on ModelNet40, ShapeNet, and the real-world dataset ScanObjectNN, demonstrated that the method achieves a great trade-off between efficiency, rotation, and accuracy.
arXiv Detail & Related papers (2022-09-13T12:12:19Z) - ThreshNet: An Efficient DenseNet using Threshold Mechanism to Reduce
Connections [1.2542322096299672]
We propose a new network architecture using threshold mechanism to further optimize the method of connections.
ThreshNet achieves up to 60% reduction in inference time compared to DenseNet, and up to 35% faster training speed and 20% reduction in error rate.
arXiv Detail & Related papers (2022-01-09T13:52:16Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - MutualNet: Adaptive ConvNet via Mutual Learning from Different Model
Configurations [51.85020143716815]
We propose MutualNet to train a single network that can run at a diverse set of resource constraints.
Our method trains a cohort of model configurations with various network widths and input resolutions.
MutualNet is a general training methodology that can be applied to various network structures.
arXiv Detail & Related papers (2021-05-14T22:30:13Z) - PocketNet: A Smaller Neural Network for 3D Medical Image Segmentation [0.0]
We derive a new CNN architecture called PocketNet that achieves comparable segmentation results to conventional CNNs while using less than 3% of the number of parameters.
We show that PocketNet achieves comparable segmentation results to conventional CNNs while using less than 3% of the number of parameters.
arXiv Detail & Related papers (2021-04-21T20:10:30Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z) - Is Each Layer Non-trivial in CNN? [11.854634156817642]
Convolutional neural network (CNN) models have achieved great success in many fields.
With the advent of ResNet, networks used in practice are getting deeper and wider.
We trained a network on the training set, then we replace the network convolution kernels with zeros and test the result models on the test set.
arXiv Detail & Related papers (2020-09-09T02:17:49Z) - Adjoined Networks: A Training Paradigm with Applications to Network
Compression [3.995047443480282]
We introduce Adjoined Networks, or AN, a learning paradigm that trains both the original base network and the smaller compressed network together.
Using ResNet-50 as the base network, AN achieves 71.8% top-1 accuracy with only 1.8M parameters and 1.6 GFLOPs on the ImageNet data-set.
We propose Differentiable Adjoined Networks (DAN), a training paradigm that augments AN by using neural architecture search to jointly learn both the width and the weights for each layer of the smaller network.
arXiv Detail & Related papers (2020-06-10T02:48:16Z) - Improved Residual Networks for Image and Video Recognition [98.10703825716142]
Residual networks (ResNets) represent a powerful type of convolutional neural network (CNN) architecture.
We show consistent improvements in accuracy and learning convergence over the baseline.
Our proposed approach allows us to train extremely deep networks, while the baseline shows severe optimization issues.
arXiv Detail & Related papers (2020-04-10T11:09:50Z) - ReActNet: Towards Precise Binary Neural Network with Generalized
Activation Functions [76.05981545084738]
We propose several ideas for enhancing a binary network to close its accuracy gap from real-valued networks without incurring any additional computational cost.
We first construct a baseline network by modifying and binarizing a compact real-valued network with parameter-free shortcuts.
We show that the proposed ReActNet outperforms all the state-of-the-arts by a large margin.
arXiv Detail & Related papers (2020-03-07T02:12:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.