Extended Batch Normalization
- URL: http://arxiv.org/abs/2003.05569v1
- Date: Thu, 12 Mar 2020 01:53:15 GMT
- Title: Extended Batch Normalization
- Authors: Chunjie Luo, Jianfeng Zhan, Lei Wang, Wanling Gao
- Abstract summary: Batch normalization (BN) has become a standard technique for training the modern deep networks.
In this paper, we propose a simple but effective method, called extended batch normalization (EBN)
Experiments show that extended batch normalization alleviates the problem of batch normalization with small batch size while achieving close performances to batch normalization with large batch size.
- Score: 3.377000738091241
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Batch normalization (BN) has become a standard technique for training the
modern deep networks. However, its effectiveness diminishes when the batch size
becomes smaller, since the batch statistics estimation becomes inaccurate. That
hinders batch normalization's usage for 1) training larger model which requires
small batches constrained by memory consumption, 2) training on mobile or
embedded devices of which the memory resource is limited. In this paper, we
propose a simple but effective method, called extended batch normalization
(EBN). For NCHW format feature maps, extended batch normalization computes the
mean along the (N, H, W) dimensions, as the same as batch normalization, to
maintain the advantage of batch normalization. To alleviate the problem caused
by small batch size, extended batch normalization computes the standard
deviation along the (N, C, H, W) dimensions, thus enlarges the number of
samples from which the standard deviation is computed. We compare extended
batch normalization with batch normalization and group normalization on the
datasets of MNIST, CIFAR-10/100, STL-10, and ImageNet, respectively. The
experiments show that extended batch normalization alleviates the problem of
batch normalization with small batch size while achieving close performances to
batch normalization with large batch size.
Related papers
- Exploring the Efficacy of Group-Normalization in Deep Learning Models for Alzheimer's Disease Classification [2.6447365674762273]
Group Normalization is an easy alternative to Batch Normalization.
GN achieves a very low error rate of 10.6% compared to Batch Normalization.
arXiv Detail & Related papers (2024-04-01T06:10:11Z) - BCN: Batch Channel Normalization for Image Classification [13.262032378453073]
This paper presents a novel normalization technique called Batch Channel Normalization (BCN)
As a basic block, BCN can be easily integrated into existing models for various applications in the field of computer vision.
arXiv Detail & Related papers (2023-12-01T14:01:48Z) - The Implicit Bias of Batch Normalization in Linear Models and Two-layer
Linear Convolutional Neural Networks [117.93273337740442]
We show that gradient descent converges to a uniform margin classifier on the training data with an $exp(-Omega(log2 t))$ convergence rate.
We also show that batch normalization has an implicit bias towards a patch-wise uniform margin.
arXiv Detail & Related papers (2023-06-20T16:58:00Z) - Batch Layer Normalization, A new normalization layer for CNNs and RNN [0.0]
This study introduces a new normalization layer termed Batch Layer Normalization (BLN)
As a combined version of batch and layer normalization, BLN adaptively puts appropriate weight on mini-batch and feature normalization based on the inverse size of mini-batches.
Test results indicate the application potential of BLN and its faster convergence than batch normalization and layer normalization in both Convolutional and Recurrent Neural Networks.
arXiv Detail & Related papers (2022-09-19T10:12:51Z) - On Batch Normalisation for Approximate Bayesian Inference [102.94525205971873]
We show that batch-normalisation does not affect the optimum of the evidence lower bound (ELBO)
We also study the Monte Carlo Batch Normalisation (MCBN) algorithm, proposed as an approximate inference technique parallel to MC Dropout.
arXiv Detail & Related papers (2020-12-24T12:40:11Z) - Batch Group Normalization [45.03388237812212]
Batch Normalization (BN) performs well at medium and large batch sizes.
BN saturates at small/extreme large batch sizes due to noisy/confused statistic calculation.
BGN is proposed to solve the noisy/confused statistic calculation of BN at small/extreme large batch sizes.
arXiv Detail & Related papers (2020-12-04T18:57:52Z) - Double Forward Propagation for Memorized Batch Normalization [68.34268180871416]
Batch Normalization (BN) has been a standard component in designing deep neural networks (DNNs)
We propose a memorized batch normalization (MBN) which considers multiple recent batches to obtain more accurate and robust statistics.
Compared to related methods, the proposed MBN exhibits consistent behaviors in both training and inference.
arXiv Detail & Related papers (2020-10-10T08:48:41Z) - Cross-Iteration Batch Normalization [67.83430009388678]
We present Cross-It Batch Normalization (CBN), in which examples from multiple recent iterations are jointly utilized to enhance estimation quality.
CBN is found to outperform the original batch normalization and a direct calculation of statistics over previous iterations without the proposed compensation technique.
arXiv Detail & Related papers (2020-02-13T18:52:57Z) - Towards Stabilizing Batch Statistics in Backward Propagation of Batch
Normalization [126.6252371899064]
Moving Average Batch Normalization (MABN) is a novel normalization method.
We show that MABN can completely restore the performance of vanilla BN in small batch cases.
Our experiments demonstrate the effectiveness of MABN in multiple computer vision tasks including ImageNet and COCO.
arXiv Detail & Related papers (2020-01-19T14:41:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.