Time Matters in Using Data Augmentation for Vision-based Deep
Reinforcement Learning
- URL: http://arxiv.org/abs/2102.08581v1
- Date: Wed, 17 Feb 2021 05:22:34 GMT
- Title: Time Matters in Using Data Augmentation for Vision-based Deep
Reinforcement Learning
- Authors: Byungchan Ko and Jungseul Ok
- Abstract summary: The timing of using augmentation, which is, in turn, critical depending on tasks to be solved in training and testing.
If the regularization imposed by augmentation is helpful only in testing, it is better to procrastinate the augmentation after training than to use it during training in terms of sample and computation complexity.
- Score: 4.921588282642753
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data augmentation technique from computer vision has been widely considered
as a regularization method to improve data efficiency and generalization
performance in vision-based reinforcement learning. We variate the timing of
using augmentation, which is, in turn, critical depending on tasks to be solved
in training and testing. According to our experiments on Open AI Procgen
Benchmark, if the regularization imposed by augmentation is helpful only in
testing, it is better to procrastinate the augmentation after training than to
use it during training in terms of sample and computation complexity. We note
that some of such augmentations can disturb the training process. Conversely,
an augmentation providing regularization useful in training needs to be used
during the whole training period to fully utilize its benefit in terms of not
only generalization but also data efficiency. These phenomena suggest a useful
timing control of data augmentation in reinforcement learning.
Related papers
- Normalization and effective learning rates in reinforcement learning [52.59508428613934]
Normalization layers have recently experienced a renaissance in the deep reinforcement learning and continual learning literature.
We show that normalization brings with it a subtle but important side effect: an equivalence between growth in the norm of the network parameters and decay in the effective learning rate.
We propose to make the learning rate schedule explicit with a simple re- parameterization which we call Normalize-and-Project.
arXiv Detail & Related papers (2024-07-01T20:58:01Z) - A Recipe for Unbounded Data Augmentation in Visual Reinforcement Learning [12.889687274108248]
A Q-learning algorithm is prone to overfitting and training instabilities when trained from visual observations.
We propose a generalized recipe, SADA, that works with wider varieties of augmentations.
We find that our method, SADA, greatly improves training stability and generalization of RL agents across a diverse set of augmentations.
arXiv Detail & Related papers (2024-05-27T17:58:23Z) - Data Augmentation for Traffic Classification [54.92823760790628]
Data Augmentation (DA) is a technique widely adopted in Computer Vision (CV) and Natural Language Processing (NLP) tasks.
DA has struggled to gain traction in networking contexts, particularly in Traffic Classification (TC) tasks.
arXiv Detail & Related papers (2024-01-19T15:25:09Z) - Incorporating Supervised Domain Generalization into Data Augmentation [4.14360329494344]
We propose a method, contrastive semantic alignment(CSA) loss, to improve robustness and training efficiency of data augmentation.
Experiments on the CIFAR-100 and CUB datasets show that the proposed method improves the robustness and training efficiency of typical data augmentations.
arXiv Detail & Related papers (2023-10-02T09:20:12Z) - Tied-Augment: Controlling Representation Similarity Improves Data
Augmentation [18.446051824487792]
We propose a framework called Tied-Augment to improve data augmentation in a wide range of applications.
Tied-Augment can improve state-of-the-art methods from data augmentation (e.g. RandAugment, mixup), optimization (e.g. SAM), and semi-supervised learning (e.g. FixMatch)
arXiv Detail & Related papers (2023-05-22T22:23:40Z) - Time Series Contrastive Learning with Information-Aware Augmentations [57.45139904366001]
A key component of contrastive learning is to select appropriate augmentations imposing some priors to construct feasible positive samples.
How to find the desired augmentations of time series data that are meaningful for given contrastive learning tasks and datasets remains an open question.
We propose a new contrastive learning approach with information-aware augmentations, InfoTS, that adaptively selects optimal augmentations for time series representation learning.
arXiv Detail & Related papers (2023-03-21T15:02:50Z) - Efficient and Effective Augmentation Strategy for Adversarial Training [48.735220353660324]
Adversarial training of Deep Neural Networks is known to be significantly more data-hungry than standard training.
We propose Diverse Augmentation-based Joint Adversarial Training (DAJAT) to use data augmentations effectively in adversarial training.
arXiv Detail & Related papers (2022-10-27T10:59:55Z) - Automatic Data Augmentation via Invariance-Constrained Learning [94.27081585149836]
Underlying data structures are often exploited to improve the solution of learning tasks.
Data augmentation induces these symmetries during training by applying multiple transformations to the input data.
This work tackles these issues by automatically adapting the data augmentation while solving the learning task.
arXiv Detail & Related papers (2022-09-29T18:11:01Z) - BERT WEAVER: Using WEight AVERaging to enable lifelong learning for
transformer-based models in biomedical semantic search engines [49.75878234192369]
We present WEAVER, a simple, yet efficient post-processing method that infuses old knowledge into the new model.
We show that applying WEAVER in a sequential manner results in similar word embedding distributions as doing a combined training on all data at once.
arXiv Detail & Related papers (2022-02-21T10:34:41Z) - Training Efficiency and Robustness in Deep Learning [2.6451769337566406]
We study approaches to improve the training efficiency and robustness of deep learning models.
We find that prioritizing learning on more informative training data increases convergence speed and improves generalization performance on test data.
We show that a redundancy-aware modification to the sampling of training data improves the training speed and develops an efficient method for detecting the diversity of training signal.
arXiv Detail & Related papers (2021-12-02T17:11:33Z) - Learning Data Augmentation with Online Bilevel Optimization for Image
Classification [14.488360021440448]
We propose an efficient approach to automatically train a network that learns an effective distribution of transformations to improve its generalization.
We show that our joint training method produces an image classification accuracy comparable to or better than carefully hand-crafted data augmentation.
arXiv Detail & Related papers (2020-06-25T21:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.