Training and Inference for Integer-Based Semantic Segmentation Network
- URL: http://arxiv.org/abs/2011.14504v1
- Date: Mon, 30 Nov 2020 02:07:07 GMT
- Title: Training and Inference for Integer-Based Semantic Segmentation Network
- Authors: Jiayi Yang, Lei Deng, Yukuan Yang, Yuan Xie, Guoqi Li
- Abstract summary: We propose a new quantization framework for training and inference of semantic segmentation networks.
Our framework is evaluated on mainstream semantic segmentation networks like FCN-VGG16 and DeepLabv3-ResNet50.
- Score: 18.457074855823315
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic segmentation has been a major topic in research and industry in
recent years. However, due to the computation complexity of pixel-wise
prediction and backpropagation algorithm, semantic segmentation has been
demanding in computation resources, resulting in slow training and inference
speed and large storage space to store models. Existing schemes that speed up
segmentation network change the network structure and come with noticeable
accuracy degradation. However, neural network quantization can be used to
reduce computation load while maintaining comparable accuracy and original
network structure. Semantic segmentation networks are different from
traditional deep convolutional neural networks (DCNNs) in many ways, and this
topic has not been thoroughly explored in existing works. In this paper, we
propose a new quantization framework for training and inference of segmentation
networks, where parameters and operations are constrained to 8-bit
integer-based values for the first time. Full quantization of the data flow and
the removal of square and root operations in batch normalization give our
framework the ability to perform inference on fixed-point devices. Our proposed
framework is evaluated on mainstream semantic segmentation networks like
FCN-VGG16 and DeepLabv3-ResNet50, achieving comparable accuracy against
floating-point framework on ADE20K dataset and PASCAL VOC 2012 dataset.
Related papers
- A Generalization of Continuous Relaxation in Structured Pruning [0.3277163122167434]
Trends indicate that deeper and larger neural networks with an increasing number of parameters achieve higher accuracy than smaller neural networks.
We generalize structured pruning with algorithms for network augmentation, pruning, sub-network collapse and removal.
The resulting CNN executes efficiently on GPU hardware without computationally expensive sparse matrix operations.
arXiv Detail & Related papers (2023-08-28T14:19:13Z) - Receptive Field-based Segmentation for Distributed CNN Inference
Acceleration in Collaborative Edge Computing [93.67044879636093]
We study inference acceleration using distributed convolutional neural networks (CNNs) in collaborative edge computing network.
We propose a novel collaborative edge computing using fused-layer parallelization to partition a CNN model into multiple blocks of convolutional layers.
arXiv Detail & Related papers (2022-07-22T18:38:11Z) - Multi-Exit Semantic Segmentation Networks [78.44441236864057]
We propose a framework for converting state-of-the-art segmentation models to MESS networks.
specially trained CNNs that employ parametrised early exits along their depth to save during inference on easier samples.
We co-optimise the number, placement and architecture of the attached segmentation heads, along with the exit policy, to adapt to the device capabilities and application-specific requirements.
arXiv Detail & Related papers (2021-06-07T11:37:03Z) - ABCNet: Attentive Bilateral Contextual Network for Efficient Semantic
Segmentation of Fine-Resolution Remote Sensing Images [5.753245638190626]
Semantic segmentation of remotely sensed images plays a crucial role in precision agriculture, environmental protection, and economic assessment.
Due to the complicated information caused by the increased spatial resolution, state-of-the-art deep learning algorithms normally utilize complex network architectures for segmentation.
We propose an Attentive Bilateral Contextual Network (ABCNet), a convolutional neural network (CNN) with double branches, with prominently lower computational consumptions compared to the cutting-edge algorithms.
arXiv Detail & Related papers (2021-02-04T10:43:08Z) - ItNet: iterative neural networks with small graphs for accurate and
efficient anytime prediction [1.52292571922932]
In this study, we introduce a class of network models that have a small memory footprint in terms of their computational graphs.
We show state-of-the-art results for semantic segmentation on the CamVid and Cityscapes datasets.
arXiv Detail & Related papers (2021-01-21T15:56:29Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Encoding the latent posterior of Bayesian Neural Networks for
uncertainty quantification [10.727102755903616]
We aim for efficient deep BNNs amenable to complex computer vision architectures.
We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer.
Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles.
arXiv Detail & Related papers (2020-12-04T19:50:09Z) - Efficient Integer-Arithmetic-Only Convolutional Neural Networks [87.01739569518513]
We replace conventional ReLU with Bounded ReLU and find that the decline is due to activation quantization.
Our integer networks achieve equivalent performance as the corresponding FPN networks, but have only 1/4 memory cost and run 2x faster on modern GPU.
arXiv Detail & Related papers (2020-06-21T08:23:03Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.