Data Augmentation via Mixed Class Interpolation using Cycle-Consistent
Generative Adversarial Networks Applied to Cross-Domain Imagery
- URL: http://arxiv.org/abs/2005.02436v2
- Date: Fri, 1 Jan 2021 22:29:13 GMT
- Title: Data Augmentation via Mixed Class Interpolation using Cycle-Consistent
Generative Adversarial Networks Applied to Cross-Domain Imagery
- Authors: Hiroshi Sasaki, Chris G. Willcocks, Toby P. Breckon
- Abstract summary: Machine learning driven object detection and classification within non-visible imagery has an important role in many fields.
However, such applications often suffer due to the limited quantity and variety of non-visible spectral domain imagery.
This paper proposes and evaluates a novel data augmentation approach that leverages the more readily available visible-band imagery.
- Score: 16.870604081967866
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning driven object detection and classification within
non-visible imagery has an important role in many fields such as night vision,
all-weather surveillance and aviation security. However, such applications
often suffer due to the limited quantity and variety of non-visible spectral
domain imagery, in contrast to the high data availability of visible-band
imagery that readily enables contemporary deep learning driven detection and
classification approaches. To address this problem, this paper proposes and
evaluates a novel data augmentation approach that leverages the more readily
available visible-band imagery via a generative domain transfer model. The
model can synthesise large volumes of non-visible domain imagery by
image-to-image (I2I) translation from the visible image domain. Furthermore, we
show that the generation of interpolated mixed class (non-visible domain) image
examples via our novel Conditional CycleGAN Mixup Augmentation (C2GMA)
methodology can lead to a significant improvement in the quality of non-visible
domain classification tasks that otherwise suffer due to limited data
availability. Focusing on classification within the Synthetic Aperture Radar
(SAR) domain, our approach is evaluated on a variation of the Statoil/C-CORE
Iceberg Classifier Challenge dataset and achieves 75.4% accuracy, demonstrating
a significant improvement when compared against traditional data augmentation
strategies (Rotation, Mixup, and MixCycleGAN).
Related papers
- Unifying Visual and Semantic Feature Spaces with Diffusion Models for Enhanced Cross-Modal Alignment [20.902935570581207]
We introduce a Multimodal Alignment and Reconstruction Network (MARNet) to enhance the model's resistance to visual noise.
MARNet includes a cross-modal diffusion reconstruction module for smoothly and stably blending information across different domains.
Experiments conducted on two benchmark datasets, Vireo-Food172 and Ingredient-101, demonstrate that MARNet effectively improves the quality of image information extracted by the model.
arXiv Detail & Related papers (2024-07-26T16:30:18Z) - DA-HFNet: Progressive Fine-Grained Forgery Image Detection and Localization Based on Dual Attention [12.36906630199689]
We construct a DA-HFNet forged image dataset guided by text or image-assisted GAN and Diffusion model.
Our goal is to utilize a hierarchical progressive network to capture forged artifacts at different scales for detection and localization.
arXiv Detail & Related papers (2024-06-03T16:13:33Z) - Enhance Image Classification via Inter-Class Image Mixup with Diffusion Model [80.61157097223058]
A prevalent strategy to bolster image classification performance is through augmenting the training set with synthetic images generated by T2I models.
In this study, we scrutinize the shortcomings of both current generative and conventional data augmentation techniques.
We introduce an innovative inter-class data augmentation method known as Diff-Mix, which enriches the dataset by performing image translations between classes.
arXiv Detail & Related papers (2024-03-28T17:23:45Z) - SatSynth: Augmenting Image-Mask Pairs through Diffusion Models for Aerial Semantic Segmentation [69.42764583465508]
We explore the potential of generative image diffusion to address the scarcity of annotated data in earth observation tasks.
To the best of our knowledge, we are the first to generate both images and corresponding masks for satellite segmentation.
arXiv Detail & Related papers (2024-03-25T10:30:22Z) - GenFace: A Large-Scale Fine-Grained Face Forgery Benchmark and Cross Appearance-Edge Learning [50.7702397913573]
The rapid advancement of photorealistic generators has reached a critical juncture where the discrepancy between authentic and manipulated images is increasingly indistinguishable.
Although there have been a number of publicly available face forgery datasets, the forgery faces are mostly generated using GAN-based synthesis technology.
We propose a large-scale, diverse, and fine-grained high-fidelity dataset, namely GenFace, to facilitate the advancement of deepfake detection.
arXiv Detail & Related papers (2024-02-03T03:13:50Z) - Improving Anomaly Segmentation with Multi-Granularity Cross-Domain
Alignment [17.086123737443714]
Anomaly segmentation plays a pivotal role in identifying atypical objects in images, crucial for hazard detection in autonomous driving systems.
While existing methods demonstrate noteworthy results on synthetic data, they often fail to consider the disparity between synthetic and real-world data domains.
We introduce the Multi-Granularity Cross-Domain Alignment framework, tailored to harmonize features across domains at both the scene and individual sample levels.
arXiv Detail & Related papers (2023-08-16T22:54:49Z) - Improved Benthic Classification using Resolution Scaling and SymmNet
Unsupervised Domain Adaptation [8.35780131268962]
We adapt the SymmNet state-of-the-art Unsupervised Domain Adaptation method with an efficient bilinear pooling layer and image scaling to normalise spatial resolution.
The results show that generic domain adaptation can be enhanced to produce a significant increase in accuracy for images from an AUV survey that differs from the training images.
arXiv Detail & Related papers (2023-03-20T09:33:47Z) - Unsupervised Domain Transfer with Conditional Invertible Neural Networks [83.90291882730925]
We propose a domain transfer approach based on conditional invertible neural networks (cINNs)
Our method inherently guarantees cycle consistency through its invertible architecture, and network training can efficiently be conducted with maximum likelihood.
Our method enables the generation of realistic spectral data and outperforms the state of the art on two downstream classification tasks.
arXiv Detail & Related papers (2023-03-17T18:00:27Z) - Hierarchical Forgery Classifier On Multi-modality Face Forgery Clues [61.37306431455152]
We propose a novel Hierarchical Forgery for Multi-modality Face Forgery Detection (HFC-MFFD)
The HFC-MFFD learns robust patches-based hybrid representation to enhance forgery authentication in multiple-modality scenarios.
The specific hierarchical face forgery is proposed to alleviate the class imbalance problem and further boost detection performance.
arXiv Detail & Related papers (2022-12-30T10:54:29Z) - Multi-Scale Multi-Target Domain Adaptation for Angle Closure
Classification [50.658613573816254]
We propose a novel Multi-scale Multi-target Domain Adversarial Network (M2DAN) for angle closure classification.
Based on these domain-invariant features at different scales, the deep model trained on the source domain is able to classify angle closure on multiple target domains.
arXiv Detail & Related papers (2022-08-25T15:27:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.