Domain Translation via Latent Space Mapping
- URL: http://arxiv.org/abs/2212.03361v1
- Date: Tue, 6 Dec 2022 23:09:40 GMT
- Title: Domain Translation via Latent Space Mapping
- Authors: Tsiry Mayet and Simon Bernard and Clement Chatelain and Romain Herault
- Abstract summary: We introduce a new unified framework called Latent Space Mapping (model)
Unlike existing approaches, we propose to further regularize each latent space using available domains by learning each dependency between pairs of domains.
- Score: 1.1470070927586016
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In this paper, we investigate the problem of multi-domain translation: given
an element $a$ of domain $A$, we would like to generate a corresponding $b$
sample in another domain $B$, and vice versa. Acquiring supervision in multiple
domains can be a tedious task, also we propose to learn this translation from
one domain to another when supervision is available as a pair $(a,b)\sim
A\times B$ and leveraging possible unpaired data when only $a\sim A$ or only
$b\sim B$ is available. We introduce a new unified framework called Latent
Space Mapping (\model) that exploits the manifold assumption in order to learn,
from each domain, a latent space. Unlike existing approaches, we propose to
further regularize each latent space using available domains by learning each
dependency between pairs of domains. We evaluate our approach in three tasks
performing i) synthetic dataset with image translation, ii) real-world task of
semantic segmentation for medical images, and iii) real-world task of facial
landmark detection.
Related papers
- Universal Multi-Domain Translation via Diffusion Routers [23.85537575452933]
Multi-domain translation (MDT) aims to learn translations between multiple domains, yet existing approaches either require fully aligneds or can only handle domain pairs seen in training.<n>We introduce universal MDT (UMDT), a generalization of MDT that seeks to translate between any pair of $K$ domains using only $K-1$ paired with a central domain.<n>We propose Diffusion Router (DR), a unified diffusion-based framework that models all central$leftarrow$non-central translations with a single noise predictor.
arXiv Detail & Related papers (2025-09-26T07:32:43Z) - A Unified Data Augmentation Framework for Low-Resource Multi-Domain Dialogue Generation [52.0964459842176]
Current state-of-the-art dialogue systems heavily rely on extensive training datasets.
We propose a novel data textbfAugmentation framework for textbfMulti-textbfDomain textbfDialogue textbfGeneration, referred to as textbfAMD$2$G.
The AMD$2$G framework consists of a data augmentation process and a two-stage training approach: domain-agnostic training and domain adaptation training.
arXiv Detail & Related papers (2024-06-14T09:52:27Z) - Multiple Noises in Diffusion Model for Semi-Supervised Multi-Domain
Translation [0.6554326244334868]
This paper proposes Multi-Domain Diffusion (MDD), a conditional diffusion framework for multi-domain translation in a semi-supervised context.
Unlike previous methods, MDD does not require defining input and output domains, allowing translation between any partition of domains within a set.
arXiv Detail & Related papers (2023-09-25T15:31:16Z) - Domain-Scalable Unpaired Image Translation via Latent Space Anchoring [88.7642967393508]
Unpaired image-to-image translation (UNIT) aims to map images between two visual domains without paired training data.
We propose a new domain-scalable UNIT method, termed as latent space anchoring.
Our method anchors images of different domains to the same latent space of frozen GANs by learning lightweight encoder and regressor models.
In the inference phase, the learned encoders and decoders of different domains can be arbitrarily combined to translate images between any two domains without fine-tuning.
arXiv Detail & Related papers (2023-06-26T17:50:02Z) - Topic-Guided Sampling For Data-Efficient Multi-Domain Stance Detection [44.06173809190896]
Stance Detection is concerned with identifying the attitudes expressed by an author towards a target of interest.
This task spans a variety of domains ranging from social media opinion identification to detecting the stance for a legal claim.
We present a topic-guided diversity sampling technique and a contrastive objective that is used for fine-tuning a stance.
arXiv Detail & Related papers (2023-06-01T15:00:39Z) - Using Language to Extend to Unseen Domains [81.37175826824625]
It is expensive to collect training data for every possible domain that a vision model may encounter when deployed.
We consider how simply verbalizing the training domain as well as domains we want to extend to but do not have data for can improve robustness.
Using a multimodal model with a joint image and language embedding space, our method LADS learns a transformation of the image embeddings from the training domain to each unseen test domain.
arXiv Detail & Related papers (2022-10-18T01:14:02Z) - WUDA: Unsupervised Domain Adaptation Based on Weak Source Domain Labels [5.718326013810649]
Unsupervised domain adaptation (UDA) for semantic segmentation addresses the cross-domain problem with fine source domain labels.
This paper defines a new task: unsupervised domain adaptation based on weak source domain labels.
arXiv Detail & Related papers (2022-10-05T08:28:57Z) - Multi-Modal Cross-Domain Alignment Network for Video Moment Retrieval [55.122020263319634]
Video moment retrieval (VMR) aims to localize the target moment from an untrimmed video according to a given language query.
In this paper, we focus on a novel task: cross-domain VMR, where fully-annotated datasets are available in one domain but the domain of interest only contains unannotated datasets.
We propose a novel Multi-Modal Cross-Domain Alignment network to transfer the annotation knowledge from the source domain to the target domain.
arXiv Detail & Related papers (2022-09-23T12:58:20Z) - Dynamic Instance Domain Adaptation [109.53575039217094]
Most studies on unsupervised domain adaptation assume that each domain's training samples come with domain labels.
We develop a dynamic neural network with adaptive convolutional kernels to generate instance-adaptive residuals to adapt domain-agnostic deep features to each individual instance.
Our model, dubbed DIDA-Net, achieves state-of-the-art performance on several commonly used single-source and multi-source UDA datasets.
arXiv Detail & Related papers (2022-03-09T20:05:54Z) - GMM-UNIT: Unsupervised Multi-Domain and Multi-Modal Image-to-Image
Translation via Attribute Gaussian Mixture Modeling [66.50914391679375]
Unsupervised image-to-image translation (UNIT) aims at learning a mapping between several visual domains by using unpaired training images.
Recent studies have shown remarkable success for multiple domains but they suffer from two main limitations.
We propose a method named GMM-UNIT, which is based on a content-attribute disentangled representation where the space is fitted with a GMM.
arXiv Detail & Related papers (2020-03-15T10:18:56Z) - LiDARNet: A Boundary-Aware Domain Adaptation Model for Point Cloud
Semantic Segmentation [18.211513930388417]
We present a boundary-aware domain adaptation model for LiDAR scan full-scene semantic segmentation (LiDARNet)
Our model can extract both the domain private features and the domain shared features with a two-branch structure.
Using our approach, we can get a single projection-based LiDAR full-scene semantic segmentation model working on both domains.
arXiv Detail & Related papers (2020-03-02T20:18:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.