Exact Feature Distribution Matching for Arbitrary Style Transfer and
Domain Generalization
- URL: http://arxiv.org/abs/2203.07740v1
- Date: Tue, 15 Mar 2022 09:18:14 GMT
- Title: Exact Feature Distribution Matching for Arbitrary Style Transfer and
Domain Generalization
- Authors: Yabin Zhang, Minghan Li, Ruihuang Li, Kui Jia, Lei Zhang
- Abstract summary: We propose to perform Exact Feature Distribution Matching (EFDM) by exactly matching the empirical Cumulative Distribution Functions (eCDFs) of image features.
A fast EHM algorithm, named Sort-Matching, is employed to perform EFDM in a plug-and-play manner with minimal cost.
The effectiveness of our proposed EFDM method is verified on a variety of AST and DG tasks, demonstrating new state-of-the-art results.
- Score: 43.19170120544387
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Arbitrary style transfer (AST) and domain generalization (DG) are important
yet challenging visual learning tasks, which can be cast as a feature
distribution matching problem. With the assumption of Gaussian feature
distribution, conventional feature distribution matching methods usually match
the mean and standard deviation of features. However, the feature distributions
of real-world data are usually much more complicated than Gaussian, which
cannot be accurately matched by using only the first-order and second-order
statistics, while it is computationally prohibitive to use high-order
statistics for distribution matching. In this work, we, for the first time to
our best knowledge, propose to perform Exact Feature Distribution Matching
(EFDM) by exactly matching the empirical Cumulative Distribution Functions
(eCDFs) of image features, which could be implemented by applying the Exact
Histogram Matching (EHM) in the image feature space. Particularly, a fast EHM
algorithm, named Sort-Matching, is employed to perform EFDM in a plug-and-play
manner with minimal cost. The effectiveness of our proposed EFDM method is
verified on a variety of AST and DG tasks, demonstrating new state-of-the-art
results. Codes are available at https://github.com/YBZh/EFDM.
Related papers
- Efficient Distribution Matching of Representations via Noise-Injected Deep InfoMax [73.03684002513218]
We enhance Deep InfoMax (DIM) to enable automatic matching of learned representations to a selected prior distribution.
We show that such modification allows for learning uniformly and normally distributed representations.
The results indicate a moderate trade-off between the performance on the downstream tasks and quality of DM.
arXiv Detail & Related papers (2024-10-09T15:40:04Z) - Open Domain Generalization with a Single Network by Regularization
Exploiting Pre-trained Features [37.518025833882334]
Open Domain Generalization (ODG) is a challenging task as it deals with distribution shifts and category shifts.
Previous work has used multiple source-specific networks, which involve a high cost.
This paper proposes a method that can handle ODG using only a single network.
arXiv Detail & Related papers (2023-12-08T16:22:10Z) - RGM: A Robust Generalizable Matching Model [49.60975442871967]
We propose a deep model for sparse and dense matching, termed RGM (Robust Generalist Matching)
To narrow the gap between synthetic training samples and real-world scenarios, we build a new, large-scale dataset with sparse correspondence ground truth.
We are able to mix up various dense and sparse matching datasets, significantly improving the training diversity.
arXiv Detail & Related papers (2023-10-18T07:30:08Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - Okapi: Generalising Better by Making Statistical Matches Match [7.392460712829188]
Okapi is a simple, efficient, and general method for robust semi-supervised learning based on online statistical matching.
Our method uses a nearest-neighbours-based matching procedure to generate cross-domain views for a consistency loss.
We show that it is in fact possible to leverage additional unlabelled data to improve upon empirical risk minimisation.
arXiv Detail & Related papers (2022-11-07T12:41:17Z) - Understanding Anomaly Detection with Deep Invertible Networks through
Hierarchies of Distributions and Features [4.25227087152716]
Convolutional networks learn similar low-level feature distributions when trained on any natural image dataset.
When the discriminative features between inliers and outliers are on a high-level, anomaly detection becomes particularly challenging.
We propose two methods to remove the negative impact of model bias and domain prior on detecting high-level differences.
arXiv Detail & Related papers (2020-06-18T20:56:14Z) - Reciprocal Adversarial Learning via Characteristic Functions [12.961770002117142]
Generative adversarial nets (GANs) have become a preferred tool for tasks involving complicated distributions.
We show how to use the characteristic function (CF) to compare the distributions rather than their moments.
We then prove an equivalence between the embedded and data domains when a reciprocal exists, where we naturally develop the GAN in an auto-encoder structure.
This efficient structure uses only two modules, together with a simple training strategy, to achieve bi-directionally generating clear images.
arXiv Detail & Related papers (2020-06-15T14:04:55Z) - Feature Quantization Improves GAN Training [126.02828112121874]
Feature Quantization (FQ) for the discriminator embeds both true and fake data samples into a shared discrete space.
Our method can be easily plugged into existing GAN models, with little computational overhead in training.
arXiv Detail & Related papers (2020-04-05T04:06:50Z) - Semi-Supervised Learning with Normalizing Flows [54.376602201489995]
FlowGMM is an end-to-end approach to generative semi supervised learning with normalizing flows.
We show promising results on a wide range of applications, including AG-News and Yahoo Answers text data.
arXiv Detail & Related papers (2019-12-30T17:36:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.