Background-Mixed Augmentation for Weakly Supervised Change Detection
- URL: http://arxiv.org/abs/2211.11478v3
- Date: Tue, 20 Jun 2023 02:35:51 GMT
- Title: Background-Mixed Augmentation for Weakly Supervised Change Detection
- Authors: Rui Huang, Ruofei Wang, Qing Guo, Jieda Wei, Yuxiang Zhang, Wei Fan,
Yang Liu
- Abstract summary: Change detection (CD) is to decouple object changes (i.e., object missing or appearing) from background changes (i.e., environment variations)
Recent deep learning-based methods develop novel network architectures or optimization strategies with paired-training examples.
We develop a novel weakly supervised training algorithm that only needs image-level labels.
- Score: 18.319961338185458
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Change detection (CD) is to decouple object changes (i.e., object missing or
appearing) from background changes (i.e., environment variations) like light
and season variations in two images captured in the same scene over a long time
span, presenting critical applications in disaster management, urban
development, etc. In particular, the endless patterns of background changes
require detectors to have a high generalization against unseen environment
variations, making this task significantly challenging. Recent deep
learning-based methods develop novel network architectures or optimization
strategies with paired-training examples, which do not handle the
generalization issue explicitly and require huge manual pixel-level annotation
efforts. In this work, for the first attempt in the CD community, we study the
generalization issue of CD from the perspective of data augmentation and
develop a novel weakly supervised training algorithm that only needs
image-level labels. Different from general augmentation techniques for
classification, we propose the background-mixed augmentation that is
specifically designed for change detection by augmenting examples under the
guidance of a set of background-changing images and letting deep CD models see
diverse environment variations. Moreover, we propose the augmented & real data
consistency loss that encourages the generalization increase significantly. Our
method as a general framework can enhance a wide range of existing deep
learning-based detectors. We conduct extensive experiments in two public
datasets and enhance four state-of-the-art methods, demonstrating the
advantages of our method. We release the code at
https://github.com/tsingqguo/bgmix.
Related papers
- Robust Scene Change Detection Using Visual Foundation Models and Cross-Attention Mechanisms [27.882122236282054]
We present a novel method for scene change detection that leverages the robust feature extraction capabilities of a visual foundational model, DINOv2.
We evaluate our approach on two benchmark datasets, VL-CMU-CD and PSCD, along with their viewpoint-varied versions.
Our experiments demonstrate significant improvements in F1-score, particularly in scenarios involving geometric changes between image pairs.
arXiv Detail & Related papers (2024-09-25T11:55:27Z) - Single-temporal Supervised Remote Change Detection for Domain Generalization [42.55492600157288]
Change detection is widely applied in remote sensing image analysis.
Existing methods require training models separately for each dataset.
We propose a multimodal contrastive learning (ChangeCLIP) based on visual-labelled pre-training for change detection domain generalization.
arXiv Detail & Related papers (2024-04-17T12:38:58Z) - Enhance Image Classification via Inter-Class Image Mixup with Diffusion Model [80.61157097223058]
A prevalent strategy to bolster image classification performance is through augmenting the training set with synthetic images generated by T2I models.
In this study, we scrutinize the shortcomings of both current generative and conventional data augmentation techniques.
We introduce an innovative inter-class data augmentation method known as Diff-Mix, which enriches the dataset by performing image translations between classes.
arXiv Detail & Related papers (2024-03-28T17:23:45Z) - Weakly Supervised Change Detection via Knowledge Distillation and
Multiscale Sigmoid Inference [26.095501974608908]
We develop a novel weakly supervised change detection technique via Knowledge Distillation and Multiscale Sigmoid Inference (KD-MSI)
Our proposed technique, with its integrated training strategy, significantly outperforms the state-of-the-art.
arXiv Detail & Related papers (2024-03-09T05:01:51Z) - Bilevel Fast Scene Adaptation for Low-Light Image Enhancement [50.639332885989255]
Enhancing images in low-light scenes is a challenging but widely concerned task in the computer vision.
Main obstacle lies in the modeling conundrum from distribution discrepancy across different scenes.
We introduce the bilevel paradigm to model the above latent correspondence.
A bilevel learning framework is constructed to endow the scene-irrelevant generality of the encoder towards diverse scenes.
arXiv Detail & Related papers (2023-06-02T08:16:21Z) - Effective Data Augmentation With Diffusion Models [65.09758931804478]
We address the lack of diversity in data augmentation with image-to-image transformations parameterized by pre-trained text-to-image diffusion models.
Our method edits images to change their semantics using an off-the-shelf diffusion model, and generalizes to novel visual concepts from a few labelled examples.
We evaluate our approach on few-shot image classification tasks, and on a real-world weed recognition task, and observe an improvement in accuracy in tested domains.
arXiv Detail & Related papers (2023-02-07T20:42:28Z) - Intra-Source Style Augmentation for Improved Domain Generalization [21.591831983223997]
We propose an intra-source style augmentation (ISSA) method to improve domain generalization in semantic segmentation.
ISSA is model-agnostic and straightforwardly applicable with CNNs and Transformers.
It is also complementary to other domain generalization techniques, e.g., it improves the recent state-of-the-art solution RobustNet by $3%$ mIoU in Cityscapes to Dark Z"urich.
arXiv Detail & Related papers (2022-10-18T21:33:25Z) - Federated Domain Generalization for Image Recognition via Cross-Client
Style Transfer [60.70102634957392]
Domain generalization (DG) has been a hot topic in image recognition, with a goal to train a general model that can perform well on unseen domains.
In this paper, we propose a novel domain generalization method for image recognition through cross-client style transfer (CCST) without exchanging data samples.
Our method outperforms recent SOTA DG methods on two DG benchmarks (PACS, OfficeHome) and a large-scale medical image dataset (Camelyon17) in the FL setting.
arXiv Detail & Related papers (2022-10-03T13:15:55Z) - dual unet:a novel siamese network for change detection with cascade
differential fusion [4.651756476458979]
We propose a novel Siamese neural network for change detection task, namely Dual-UNet.
In contrast to previous individually encoded the bitemporal images, we design an encoder differential-attention module to focus on the spatial difference relationships of pixels.
Experiments demonstrate that the proposed approach consistently outperforms the most advanced methods on popular seasonal change detection datasets.
arXiv Detail & Related papers (2022-08-12T14:24:09Z) - Weakly Supervised Change Detection Using Guided Anisotropic Difusion [97.43170678509478]
We propose original ideas that help us to leverage such datasets in the context of change detection.
First, we propose the guided anisotropic diffusion (GAD) algorithm, which improves semantic segmentation results.
We then show its potential in two weakly-supervised learning strategies tailored for change detection.
arXiv Detail & Related papers (2021-12-31T10:03:47Z) - Towards Universal Representation Learning for Deep Face Recognition [106.21744671876704]
We propose a universal representation learning framework that can deal with larger variation unseen in the given training data without leveraging target domain knowledge.
Experiments show that our method achieves top performance on general face recognition datasets such as LFW and MegaFace.
arXiv Detail & Related papers (2020-02-26T23:29:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.