Passive Batch Injection Training Technique: Boosting Network Performance
by Injecting Mini-Batches from a different Data Distribution
- URL: http://arxiv.org/abs/2006.04406v1
- Date: Mon, 8 Jun 2020 08:17:32 GMT
- Title: Passive Batch Injection Training Technique: Boosting Network Performance
by Injecting Mini-Batches from a different Data Distribution
- Authors: Pravendra Singh, Pratik Mazumder, Vinay P. Namboodiri
- Abstract summary: This work presents a novel training technique for deep neural networks that makes use of additional data from a distribution that is different from that of the original input data.
To the best of our knowledge, this is the first work that makes use of different data distribution to aid the training of convolutional neural networks (CNNs)
- Score: 39.8046809855363
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work presents a novel training technique for deep neural networks that
makes use of additional data from a distribution that is different from that of
the original input data. This technique aims to reduce overfitting and improve
the generalization performance of the network. Our proposed technique, namely
Passive Batch Injection Training Technique (PBITT), even reduces the level of
overfitting in networks that already use the standard techniques for reducing
overfitting such as $L_2$ regularization and batch normalization, resulting in
significant accuracy improvements. Passive Batch Injection Training Technique
(PBITT) introduces a few passive mini-batches into the training process that
contain data from a distribution that is different from the input data
distribution. This technique does not increase the number of parameters in the
final model and also does not increase the inference (test) time but still
improves the performance of deep CNNs. To the best of our knowledge, this is
the first work that makes use of different data distribution to aid the
training of convolutional neural networks (CNNs). We thoroughly evaluate the
proposed approach on standard architectures: VGG, ResNet, and WideResNet, and
on several popular datasets: CIFAR-10, CIFAR-100, SVHN, and ImageNet. We
observe consistent accuracy improvement by using the proposed technique. We
also show experimentally that the model trained by our technique generalizes
well to other tasks such as object detection on the MS-COCO dataset using
Faster R-CNN. We present extensive ablations to validate the proposed approach.
Our approach improves the accuracy of VGG-16 by a significant margin of 2.1%
over the CIFAR-100 dataset.
Related papers
- KAKURENBO: Adaptively Hiding Samples in Deep Neural Network Training [2.8804804517897935]
We propose a method for hiding the least-important samples during the training of deep neural networks.
We adaptively find samples to exclude in a given epoch based on their contribution to the overall learning process.
Our method can reduce total training time by up to 22% impacting accuracy only by 0.4% compared to the baseline.
arXiv Detail & Related papers (2023-10-16T06:19:29Z) - On the effectiveness of partial variance reduction in federated learning
with heterogeneous data [27.527995694042506]
We show that the diversity of the final classification layers across clients impedes the performance of the FedAvg algorithm.
Motivated by this, we propose to correct model by variance reduction only on the final layers.
We demonstrate that this significantly outperforms existing benchmarks at a similar or lower communication cost.
arXiv Detail & Related papers (2022-12-05T11:56:35Z) - TCT: Convexifying Federated Learning using Bootstrapped Neural Tangent
Kernels [141.29156234353133]
State-of-the-art convex learning methods can perform far worse than their centralized counterparts when clients have dissimilar data distributions.
We show this disparity can largely be attributed to challenges presented by non-NISTity.
We propose a Train-Convexify neural network (TCT) procedure to sidestep this issue.
arXiv Detail & Related papers (2022-07-13T16:58:22Z) - Distributed Adversarial Training to Robustify Deep Neural Networks at
Scale [100.19539096465101]
Current deep neural networks (DNNs) are vulnerable to adversarial attacks, where adversarial perturbations to the inputs can change or manipulate classification.
To defend against such attacks, an effective approach, known as adversarial training (AT), has been shown to mitigate robust training.
We propose a large-batch adversarial training framework implemented over multiple machines.
arXiv Detail & Related papers (2022-06-13T15:39:43Z) - Invariance Learning in Deep Neural Networks with Differentiable Laplace
Approximations [76.82124752950148]
We develop a convenient gradient-based method for selecting the data augmentation.
We use a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective.
arXiv Detail & Related papers (2022-02-22T02:51:11Z) - Semantic Perturbations with Normalizing Flows for Improved
Generalization [62.998818375912506]
We show that perturbations in the latent space can be used to define fully unsupervised data augmentations.
We find that our latent adversarial perturbations adaptive to the classifier throughout its training are most effective.
arXiv Detail & Related papers (2021-08-18T03:20:00Z) - Cascade Bagging for Accuracy Prediction with Few Training Samples [8.373420721376739]
We propose a novel framework to train an accuracy predictor under few training samples.
The framework consists ofdata augmentation methods and an ensemble learning algorithm.
arXiv Detail & Related papers (2021-08-12T09:10:52Z) - Training Sparse Neural Networks using Compressed Sensing [13.84396596420605]
We develop and test a novel method based on compressed sensing which combines the pruning and training into a single step.
Specifically, we utilize an adaptively weighted $ell1$ penalty on the weights during training, which we combine with a generalization of the regularized dual averaging (RDA) algorithm in order to train sparse neural networks.
arXiv Detail & Related papers (2020-08-21T19:35:54Z) - ScopeFlow: Dynamic Scene Scoping for Optical Flow [94.42139459221784]
We propose to modify the common training protocols of optical flow.
The improvement is based on observing the bias in sampling challenging data.
We find that both regularization and augmentation should decrease during the training protocol.
arXiv Detail & Related papers (2020-02-25T09:58:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.