Can we learn better with hard samples?
- URL: http://arxiv.org/abs/2304.03486v1
- Date: Fri, 7 Apr 2023 05:45:26 GMT
- Title: Can we learn better with hard samples?
- Authors: Subin Sahayam, John Zakkam, Umarani Jayaraman
- Abstract summary: A variant of the traditional algorithm has been proposed, which trains the network focusing on mini-batches with high loss.
We show that the proposed method generalizes in 26.47% less number of epochs than the traditional mini-batch method in EfficientNet-B4 on STL-10.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In deep learning, mini-batch training is commonly used to optimize network
parameters. However, the traditional mini-batch method may not learn the
under-represented samples and complex patterns in the data, leading to a longer
time for generalization. To address this problem, a variant of the traditional
algorithm has been proposed, which trains the network focusing on mini-batches
with high loss. The study evaluates the effectiveness of the proposed training
using various deep neural networks trained on three benchmark datasets
(CIFAR-10, CIFAR-100, and STL-10). The deep neural networks used in the study
are ResNet-18, ResNet-50, Efficient Net B4, EfficientNetV2-S, and
MobilenetV3-S. The experimental results showed that the proposed method can
significantly improve the test accuracy and speed up the convergence compared
to the traditional mini-batch training method. Furthermore, we introduce a
hyper-parameter delta ({\delta}) that decides how many mini-batches are
considered for training. Experiments on various values of {\delta} found that
the performance of the proposed method for smaller {\delta} values generally
results in similar test accuracy and faster generalization. We show that the
proposed method generalizes in 26.47% less number of epochs than the
traditional mini-batch method in EfficientNet-B4 on STL-10. The proposed method
also improves the test top-1 accuracy by 7.26% in ResNet-18 on CIFAR-100.
Related papers
- Just How Flexible are Neural Networks in Practice? [89.80474583606242]
It is widely believed that a neural network can fit a training set containing at least as many samples as it has parameters.
In practice, however, we only find solutions via our training procedure, including the gradient and regularizers, limiting flexibility.
arXiv Detail & Related papers (2024-06-17T12:24:45Z) - Dynamic Sparse Training via Balancing the Exploration-Exploitation
Trade-off [19.230329532065635]
Sparse training could significantly mitigate the training costs by reducing the model size.
Existing sparse training methods mainly use either random-based or greedy-based drop-and-grow strategies.
In this work, we consider the dynamic sparse training as a sparse connectivity search problem.
Experimental results show that sparse models (up to 98% sparsity) obtained by our proposed method outperform the SOTA sparse training methods.
arXiv Detail & Related papers (2022-11-30T01:22:25Z) - An Experimental Study of the Impact of Pre-training on the Pruning of a
Convolutional Neural Network [0.0]
In recent years, deep neural networks have known a wide success in various application domains.
Deep neural networks usually involve a large number of parameters, which correspond to the weights of the network.
The pruning methods notably attempt to reduce the size of the parameter set, by identifying and removing the irrelevant weights.
arXiv Detail & Related papers (2021-12-15T16:02:15Z) - Accelerating Training and Inference of Graph Neural Networks with Fast
Sampling and Pipelining [58.10436813430554]
Mini-batch training of graph neural networks (GNNs) requires a lot of computation and data movement.
We argue in favor of performing mini-batch training with neighborhood sampling in a distributed multi-GPU environment.
We present a sequence of improvements to mitigate these bottlenecks, including a performance-engineered neighborhood sampler.
We also conduct an empirical analysis that supports the use of sampling for inference, showing that test accuracies are not materially compromised.
arXiv Detail & Related papers (2021-10-16T02:41:35Z) - Concurrent Adversarial Learning for Large-Batch Training [83.55868483681748]
Adversarial learning is a natural choice for smoothing the decision surface and biasing towards a flat region.
We propose a novel Concurrent Adversarial Learning (ConAdv) method that decouples the sequential gradient computations in adversarial learning by utilizing staled parameters.
This is the first work successfully scales ResNet-50 training batch size to 96K.
arXiv Detail & Related papers (2021-06-01T04:26:02Z) - Efficient Sparse Artificial Neural Networks [11.945854832533232]
The brain, as the source of inspiration for Artificial Neural Networks (ANN), is based on a sparse structure.
This sparse structure helps the brain to consume less energy, learn easier and generalize patterns better than any other ANN.
In this paper, two evolutionary methods for adopting sparsity to ANNs are proposed.
arXiv Detail & Related papers (2021-03-13T10:03:41Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Local Critic Training for Model-Parallel Learning of Deep Neural
Networks [94.69202357137452]
We propose a novel model-parallel learning method, called local critic training.
We show that the proposed approach successfully decouples the update process of the layer groups for both convolutional neural networks (CNNs) and recurrent neural networks (RNNs)
We also show that trained networks by the proposed method can be used for structural optimization.
arXiv Detail & Related papers (2021-02-03T09:30:45Z) - Training Sparse Neural Networks using Compressed Sensing [13.84396596420605]
We develop and test a novel method based on compressed sensing which combines the pruning and training into a single step.
Specifically, we utilize an adaptively weighted $ell1$ penalty on the weights during training, which we combine with a generalization of the regularized dual averaging (RDA) algorithm in order to train sparse neural networks.
arXiv Detail & Related papers (2020-08-21T19:35:54Z) - Passive Batch Injection Training Technique: Boosting Network Performance
by Injecting Mini-Batches from a different Data Distribution [39.8046809855363]
This work presents a novel training technique for deep neural networks that makes use of additional data from a distribution that is different from that of the original input data.
To the best of our knowledge, this is the first work that makes use of different data distribution to aid the training of convolutional neural networks (CNNs)
arXiv Detail & Related papers (2020-06-08T08:17:32Z) - Large Batch Training Does Not Need Warmup [111.07680619360528]
Training deep neural networks using a large batch size has shown promising results and benefits many real-world applications.
In this paper, we propose a novel Complete Layer-wise Adaptive Rate Scaling (CLARS) algorithm for large-batch training.
Based on our analysis, we bridge the gap and illustrate the theoretical insights for three popular large-batch training techniques.
arXiv Detail & Related papers (2020-02-04T23:03:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.