Progressive Random Convolutions for Single Domain Generalization
- URL: http://arxiv.org/abs/2304.00424v1
- Date: Sun, 2 Apr 2023 01:42:51 GMT
- Title: Progressive Random Convolutions for Single Domain Generalization
- Authors: Seokeon Choi, Debasmit Das, Sungha Choi, Seunghan Yang, Hyunsin Park,
Sungrack Yun
- Abstract summary: Single domain generalization aims to train a generalizable model with only one source domain to perform well on arbitrary unseen target domains.
Image augmentation based on Random Convolutions (RandConv) enables the model to learn generalizable visual representations by distorting local textures.
We propose a Progressive Random Convolution (Pro-RandConv) method that stacks random convolution layers with a small kernel size instead of increasing the kernel size.
- Score: 23.07924668615951
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Single domain generalization aims to train a generalizable model with only
one source domain to perform well on arbitrary unseen target domains. Image
augmentation based on Random Convolutions (RandConv), consisting of one
convolution layer randomly initialized for each mini-batch, enables the model
to learn generalizable visual representations by distorting local textures
despite its simple and lightweight structure. However, RandConv has structural
limitations in that the generated image easily loses semantics as the kernel
size increases, and lacks the inherent diversity of a single convolution
operation. To solve the problem, we propose a Progressive Random Convolution
(Pro-RandConv) method that recursively stacks random convolution layers with a
small kernel size instead of increasing the kernel size. This progressive
approach can not only mitigate semantic distortions by reducing the influence
of pixels away from the center in the theoretical receptive field, but also
create more effective virtual domains by gradually increasing the style
diversity. In addition, we develop a basic random convolution layer into a
random convolution block including deformable offsets and affine transformation
to support texture and contrast diversification, both of which are also
randomly initialized. Without complex generators or adversarial learning, we
demonstrate that our simple yet effective augmentation strategy outperforms
state-of-the-art methods on single domain generalization benchmarks.
Related papers
- Padding-free Convolution based on Preservation of Differential
Characteristics of Kernels [1.3597551064547502]
We present a non-padding-based method for size-keeping convolution based on the preservation of differential characteristics of kernels.
The main idea is to make convolution over an incomplete sliding window "collapse" to a linear differential operator evaluated locally at its central pixel.
arXiv Detail & Related papers (2023-09-12T16:36:12Z) - Look at the Neighbor: Distortion-aware Unsupervised Domain Adaptation
for Panoramic Semantic Segmentation [5.352137021024213]
The aim is to tackle the domain gaps caused by the style disparities and distortion problem from the non-uniformly distributed pixels of equirectangular projection (ERP)
We propose a novel UDA framework that can effectively address the distortion problems for panoramic semantic segmentation.
arXiv Detail & Related papers (2023-08-10T10:47:12Z) - OneDConv: Generalized Convolution For Transform-Invariant Representation [76.15687106423859]
We propose a novel generalized one dimension convolutional operator (OneDConv)
It dynamically transforms the convolution kernels based on the input features in a computationally and parametrically efficient manner.
It improves the robustness and generalization of convolution without sacrificing the performance on common images.
arXiv Detail & Related papers (2022-01-15T07:44:44Z) - Adaptive Single Image Deblurring [43.02281823557039]
We propose an efficient pixel adaptive and feature attentive design for handling large blur variations within and across different images.
We also propose an effective content-aware global-local filtering module that significantly improves the performance.
arXiv Detail & Related papers (2022-01-01T10:10:19Z) - CSformer: Bridging Convolution and Transformer for Compressive Sensing [65.22377493627687]
This paper proposes a hybrid framework that integrates the advantages of leveraging detailed spatial information from CNN and the global context provided by transformer for enhanced representation learning.
The proposed approach is an end-to-end compressive image sensing method, composed of adaptive sampling and recovery.
The experimental results demonstrate the effectiveness of the dedicated transformer-based architecture for compressive sensing.
arXiv Detail & Related papers (2021-12-31T04:37:11Z) - Learning with convolution and pooling operations in kernel methods [8.528384027684192]
Recent empirical work has shown that hierarchical convolutional kernels improve the performance of kernel methods in image classification tasks.
We study the precise interplay between approximation and generalization in convolutional architectures.
Our results quantify how choosing an architecture adapted to the target function leads to a large improvement in the sample complexity.
arXiv Detail & Related papers (2021-11-16T09:00:44Z) - Mutual Affine Network for Spatially Variant Kernel Estimation in Blind
Image Super-Resolution [130.32026819172256]
Existing blind image super-resolution (SR) methods mostly assume blur kernels are spatially invariant across the whole image.
This paper proposes a mutual affine network (MANet) for spatially variant kernel estimation.
arXiv Detail & Related papers (2021-08-11T16:11:17Z) - X-volution: On the unification of convolution and self-attention [52.80459687846842]
We propose a multi-branch elementary module composed of both convolution and self-attention operation.
The proposed X-volution achieves highly competitive visual understanding improvements.
arXiv Detail & Related papers (2021-06-04T04:32:02Z) - Robust and Generalizable Visual Representation Learning via Random
Convolutions [44.62476686073595]
We show that the robustness of neural networks can be greatly improved through the use of random convolutions as data augmentation.
Our method can benefit downstream tasks by providing a more robust pretrained visual representation.
arXiv Detail & Related papers (2020-07-25T19:52:25Z) - Locally Masked Convolution for Autoregressive Models [107.4635841204146]
LMConv is a simple modification to the standard 2D convolution that allows arbitrary masks to be applied to the weights at each location in the image.
We learn an ensemble of distribution estimators that share parameters but differ in generation order, achieving improved performance on whole-image density estimation.
arXiv Detail & Related papers (2020-06-22T17:59:07Z) - FDA: Fourier Domain Adaptation for Semantic Segmentation [82.4963423086097]
We describe a simple method for unsupervised domain adaptation, whereby the discrepancy between the source and target distributions is reduced by swapping the low-frequency spectrum of one with the other.
We illustrate the method in semantic segmentation, where densely annotated images are aplenty in one domain, but difficult to obtain in another.
Our results indicate that even simple procedures can discount nuisance variability in the data that more sophisticated methods struggle to learn away.
arXiv Detail & Related papers (2020-04-11T22:20:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.