Controlling Directions Orthogonal to a Classifier
- URL: http://arxiv.org/abs/2201.11259v1
- Date: Thu, 27 Jan 2022 01:23:08 GMT
- Title: Controlling Directions Orthogonal to a Classifier
- Authors: Yilun Xu, Hao He, Tianxiao Shen, Tommi Jaakkola
- Abstract summary: We propose to identify directions invariant to a given classifier so that these directions can be controlled in tasks such as style transfer.
We present three use cases where controlling orthogonal variation is important: style transfer, domain adaptation, and fairness.
The code is available at http://github.com/Newbeeer/orthogonal_classifier.
- Score: 11.882219706353045
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose to identify directions invariant to a given classifier so that
these directions can be controlled in tasks such as style transfer. While
orthogonal decomposition is directly identifiable when the given classifier is
linear, we formally define a notion of orthogonality in the non-linear case. We
also provide a surprisingly simple method for constructing the orthogonal
classifier (a classifier utilizing directions other than those of the given
classifier). Empirically, we present three use cases where controlling
orthogonal variation is important: style transfer, domain adaptation, and
fairness. The orthogonal classifier enables desired style transfer when domains
vary in multiple aspects, improves domain adaptation with label shifts and
mitigates the unfairness as a predictor. The code is available at
http://github.com/Newbeeer/orthogonal_classifier
Related papers
- Householder Projector for Unsupervised Latent Semantics Discovery [58.92485745195358]
Householder Projector helps StyleGANs to discover more disentangled and precise semantic attributes without sacrificing image fidelity.
We integrate our projector into pre-trained StyleGAN2/StyleGAN3 and evaluate the models on several benchmarks.
arXiv Detail & Related papers (2023-07-16T11:43:04Z) - Self-Supervised Learning for Group Equivariant Neural Networks [75.62232699377877]
Group equivariant neural networks are the models whose structure is restricted to commute with the transformations on the input.
We propose two concepts for self-supervised tasks: equivariant pretext labels and invariant contrastive loss.
Experiments on standard image recognition benchmarks demonstrate that the equivariant neural networks exploit the proposed self-supervised tasks.
arXiv Detail & Related papers (2023-03-08T08:11:26Z) - On the Implicit Bias of Linear Equivariant Steerable Networks [9.539074889921935]
We study the implicit bias of gradient flow on linear equivariant steerable networks in group-invariant binary classification.
Under a unitary assumption on the input representation, we establish the equivalence between steerable networks and data augmentation.
arXiv Detail & Related papers (2023-03-07T19:37:35Z) - Adapting to Latent Subgroup Shifts via Concepts and Proxies [82.01141290360562]
We show that the optimal target predictor can be non-parametrically identified with the help of concept and proxy variables available only in the source domain.
For continuous observations, we propose a latent variable model specific to the data generation process at hand.
arXiv Detail & Related papers (2022-12-21T18:30:22Z) - A Simple Strategy to Provable Invariance via Orbit Mapping [14.127786615513978]
We propose a method to make network architectures provably invariant with respect to group actions.
In a nutshell, we intend to 'undo' any possible transformation before feeding the data into the actual network.
arXiv Detail & Related papers (2022-09-24T03:40:42Z) - Cross-domain Detection Transformer based on Spatial-aware and
Semantic-aware Token Alignment [31.759205815348658]
We propose a new method called Spatial-aware and Semantic-aware Token Alignment (SSTA) for cross-domain detection transformers.
For spatial-aware token alignment, we can extract the information from the cross-attention map (CAM) to align the distribution of tokens according to their attention to object queries.
For semantic-aware token alignment, we inject the category information into the cross-attention map and construct domain embedding to guide the learning of a multi-class discriminator.
arXiv Detail & Related papers (2022-06-01T04:13:22Z) - Shape-Pose Disentanglement using SE(3)-equivariant Vector Neurons [59.83721247071963]
We introduce an unsupervised technique for encoding point clouds into a canonical shape representation, by disentangling shape and pose.
Our encoder is stable and consistent, meaning that the shape encoding is purely pose-invariant.
The extracted rotation and translation are able to semantically align different input shapes of the same class to a common canonical pose.
arXiv Detail & Related papers (2022-04-03T21:00:44Z) - On the rate of convergence of a classifier based on a Transformer
encoder [55.41148606254641]
The rate of convergence of the misclassification probability of the classifier towards the optimal misclassification probability is analyzed.
It is shown that this classifier is able to circumvent the curse of dimensionality provided the aposteriori probability satisfies a suitable hierarchical composition model.
arXiv Detail & Related papers (2021-11-29T14:58:29Z) - Commutative Lie Group VAE for Disentanglement Learning [96.32813624341833]
We view disentanglement learning as discovering an underlying structure that equivariantly reflects the factorized variations shown in data.
A simple model named Commutative Lie Group VAE is introduced to realize the group-based disentanglement learning.
Experiments show that our model can effectively learn disentangled representations without supervision, and can achieve state-of-the-art performance without extra constraints.
arXiv Detail & Related papers (2021-06-07T07:03:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.