Learning Representational Invariances for Data-Efficient Action
Recognition
- URL: http://arxiv.org/abs/2103.16565v1
- Date: Tue, 30 Mar 2021 17:59:49 GMT
- Title: Learning Representational Invariances for Data-Efficient Action
Recognition
- Authors: Yuliang Zou, Jinwoo Choi, Qitong Wang, Jia-Bin Huang
- Abstract summary: We show that our data augmentation strategy leads to promising performance on the Kinetics-100, UCF-101, and HMDB-51 datasets.
We also validate our data augmentation strategy in the fully supervised setting and demonstrate improved performance.
- Score: 52.23716087656834
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data augmentation is a ubiquitous technique for improving image
classification when labeled data is scarce. Constraining the model predictions
to be invariant to diverse data augmentations effectively injects the desired
representational invariances to the model (e.g., invariance to photometric
variations), leading to improved accuracy. Compared to image data, the
appearance variations in videos are far more complex due to the additional
temporal dimension. Yet, data augmentation methods for videos remain
under-explored. In this paper, we investigate various data augmentation
strategies that capture different video invariances, including photometric,
geometric, temporal, and actor/scene augmentations. When integrated with
existing consistency-based semi-supervised learning frameworks, we show that
our data augmentation strategy leads to promising performance on the
Kinetics-100, UCF-101, and HMDB-51 datasets in the low-label regime. We also
validate our data augmentation strategy in the fully supervised setting and
demonstrate improved performance.
Related papers
- Data Augmentation via Latent Diffusion for Saliency Prediction [67.88936624546076]
Saliency prediction models are constrained by the limited diversity and quantity of labeled data.
We propose a novel data augmentation method for deep saliency prediction that edits natural images while preserving the complexity and variability of real-world scenes.
arXiv Detail & Related papers (2024-09-11T14:36:24Z) - A Simple Background Augmentation Method for Object Detection with Diffusion Model [53.32935683257045]
In computer vision, it is well-known that a lack of data diversity will impair model performance.
We propose a simple yet effective data augmentation approach by leveraging advancements in generative models.
Background augmentation, in particular, significantly improves the models' robustness and generalization capabilities.
arXiv Detail & Related papers (2024-08-01T07:40:00Z) - Generalization Gap in Data Augmentation: Insights from Illumination [3.470401787749558]
We investigate the differences in generalization between models trained with augmented data and those trained under real-world illumination conditions.
Results indicate that after applying various data augmentation methods, model performance has significantly improved.
Yet, a noticeable generalization gap still exists after utilizing various data augmentation methods.
arXiv Detail & Related papers (2024-04-11T07:11:43Z) - Do Generated Data Always Help Contrastive Learning? [32.58214897368031]
Contrastive Learning (CL) has emerged as one of the most successful paradigms for unsupervised visual representation learning.
With the rise of generative models, especially diffusion models, the ability to generate realistic images close to the real data distribution has been well recognized.
However, we find that the generated data (even from a good diffusion model like DDPM) may sometimes even harm contrastive learning.
arXiv Detail & Related papers (2024-03-19T05:17:47Z) - Effective Data Augmentation With Diffusion Models [65.09758931804478]
We address the lack of diversity in data augmentation with image-to-image transformations parameterized by pre-trained text-to-image diffusion models.
Our method edits images to change their semantics using an off-the-shelf diffusion model, and generalizes to novel visual concepts from a few labelled examples.
We evaluate our approach on few-shot image classification tasks, and on a real-world weed recognition task, and observe an improvement in accuracy in tested domains.
arXiv Detail & Related papers (2023-02-07T20:42:28Z) - TeachAugment: Data Augmentation Optimization Using Teacher Knowledge [11.696069523681178]
We propose a data augmentation optimization method based on the adversarial strategy called TeachAugment.
We show that TeachAugment outperforms existing methods in experiments of image classification, semantic segmentation, and unsupervised representation learning tasks.
arXiv Detail & Related papers (2022-02-25T06:22:51Z) - Negative Data Augmentation [127.28042046152954]
We show that negative data augmentation samples provide information on the support of the data distribution.
We introduce a new GAN training objective where we use NDA as an additional source of synthetic data for the discriminator.
Empirically, models trained with our method achieve improved conditional/unconditional image generation along with improved anomaly detection capabilities.
arXiv Detail & Related papers (2021-02-09T20:28:35Z) - Data Augmentation for Meta-Learning [58.47185740820304]
meta-learning algorithms sample data, query data, and tasks on each training step.
Data augmentation can be used not only to expand the number of images available per class, but also to generate entirely new classes/tasks.
Our proposed meta-specific data augmentation significantly improves the performance of meta-learners on few-shot classification benchmarks.
arXiv Detail & Related papers (2020-10-14T13:48:22Z) - On the Generalization Effects of Linear Transformations in Data
Augmentation [32.01435459892255]
Data augmentation is a powerful technique to improve performance in applications such as image and text classification tasks.
We study a family of linear transformations and study their effects on the ridge estimator in an over-parametrized linear regression setting.
We propose an augmentation scheme that searches over the space of transformations by how uncertain the model is about the transformed data.
arXiv Detail & Related papers (2020-05-02T04:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.