Feature transforms for image data augmentation
- URL: http://arxiv.org/abs/2201.09700v1
- Date: Mon, 24 Jan 2022 14:12:29 GMT
- Title: Feature transforms for image data augmentation
- Authors: Loris Nanni, Michelangelo Paci, Sheryl Brahnam and Alessandra Lumini
- Abstract summary: In image classification, many augmentation approaches utilize simple image manipulation algorithms.
In this work, we build ensembles on the data level by adding images generated by combining fourteen augmentation approaches.
Pretrained ResNet50 networks are finetuned on training sets that include images derived from each augmentation method.
- Score: 74.12025519234153
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A problem with Convolutional Neural Networks (CNNs) is that they require
large datasets to obtain adequate robustness; on small datasets, they are prone
to overfitting. Many methods have been proposed to overcome this shortcoming
with CNNs. In cases where additional samples cannot easily be collected, a
common approach is to generate more data points from existing data using an
augmentation technique. In image classification, many augmentation approaches
utilize simple image manipulation algorithms. In this work, we build ensembles
on the data level by adding images generated by combining fourteen augmentation
approaches, three of which are proposed here for the first time. These novel
methods are based on the Fourier Transform (FT), the Radon Transform (RT) and
the Discrete Cosine Transform (DCT). Pretrained ResNet50 networks are finetuned
on training sets that include images derived from each augmentation method.
These networks and several fusions are evaluated and compared across eleven
benchmarks. Results show that building ensembles on the data level by combining
different data augmentation methods produce classifiers that not only compete
competitively against the state-of-the-art but often surpass the best
approaches reported in the literature.
Related papers
- Additional Look into GAN-based Augmentation for Deep Learning COVID-19
Image Classification [57.1795052451257]
We study the dependence of the GAN-based augmentation performance on dataset size with a focus on small samples.
We train StyleGAN2-ADA with both sets and then, after validating the quality of generated images, we use trained GANs as one of the augmentations approaches in multi-class classification problems.
The GAN-based augmentation approach is found to be comparable with classical augmentation in the case of medium and large datasets but underperforms in the case of smaller datasets.
arXiv Detail & Related papers (2024-01-26T08:28:13Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - Generalizing Dataset Distillation via Deep Generative Prior [75.9031209877651]
We propose to distill an entire dataset's knowledge into a few synthetic images.
The idea is to synthesize a small number of synthetic data points that, when given to a learning algorithm as training data, result in a model approximating one trained on the original data.
We present a new optimization algorithm that distills a large number of images into a few intermediate feature vectors in the generative model's latent space.
arXiv Detail & Related papers (2023-05-02T17:59:31Z) - ScoreMix: A Scalable Augmentation Strategy for Training GANs with
Limited Data [93.06336507035486]
Generative Adversarial Networks (GANs) typically suffer from overfitting when limited training data is available.
We present ScoreMix, a novel and scalable data augmentation approach for various image synthesis tasks.
arXiv Detail & Related papers (2022-10-27T02:55:15Z) - Ensemble Augmentation for Deep Neural Networks Using 1-D Time Series
Vibration Data [0.0]
Time-series data are one of the fundamental types of raw data representation used in data-driven techniques.
Deep Neural Networks (DNNs) require huge labeled training samples to reach their optimum performance.
In this study, a data augmentation technique named ensemble augmentation is proposed to overcome this limitation.
arXiv Detail & Related papers (2021-08-06T20:04:29Z) - FeatMatch: Feature-Based Augmentation for Semi-Supervised Learning [64.32306537419498]
We propose a novel learned feature-based refinement and augmentation method that produces a varied set of complex transformations.
These transformations also use information from both within-class and across-class representations that we extract through clustering.
We demonstrate that our method is comparable to current state of art for smaller datasets while being able to scale up to larger datasets.
arXiv Detail & Related papers (2020-07-16T17:55:31Z) - Complex Wavelet SSIM based Image Data Augmentation [0.0]
We look at the MNIST handwritten dataset an image dataset used for digit recognition.
We take a detailed look into one of the most popular augmentation techniques used for this data set elastic deformation.
We propose to use a similarity measure called Complex Wavelet Structural Similarity Index Measure (CWSSIM) to selectively filter out the irrelevant data.
arXiv Detail & Related papers (2020-07-11T21:11:46Z) - FuCiTNet: Improving the generalization of deep learning networks by the
fusion of learned class-inherent transformations [1.8013893443965217]
It is widely known that very small datasets produce overfitting in Deep Neural Networks (DNNs)
This work presents a new approach, independent but complementary to the previous mentioned techniques, for improving the generalization of DNNs on very small datasets.
arXiv Detail & Related papers (2020-05-17T12:04:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.