Unsupervised Noise Adaptive Speech Enhancement by
Discriminator-Constrained Optimal Transport
- URL: http://arxiv.org/abs/2111.06316v1
- Date: Thu, 11 Nov 2021 17:15:37 GMT
- Title: Unsupervised Noise Adaptive Speech Enhancement by
Discriminator-Constrained Optimal Transport
- Authors: Hsin-Yi Lin, Huan-Hsin Tseng, Xugang Lu and Yu Tsao
- Abstract summary: This paper presents a novel discriminator-constrained optimal transport network (DOTN) that performs unsupervised domain adaptation for speech enhancement (SE)
The DOTN aims to estimate clean references of noisy speech in a target domain, by exploiting the knowledge available from the source domain.
- Score: 25.746489468835357
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a novel discriminator-constrained optimal transport
network (DOTN) that performs unsupervised domain adaptation for speech
enhancement (SE), which is an essential regression task in speech processing.
The DOTN aims to estimate clean references of noisy speech in a target domain,
by exploiting the knowledge available from the source domain. The domain shift
between training and testing data has been reported to be an obstacle to
learning problems in diverse fields. Although rich literature exists on
unsupervised domain adaptation for classification, the methods proposed,
especially in regressions, remain scarce and often depend on additional
information regarding the input data. The proposed DOTN approach tactically
fuses the optimal transport (OT) theory from mathematical analysis with
generative adversarial frameworks, to help evaluate continuous labels in the
target domain. The experimental results on two SE tasks demonstrate that by
extending the classical OT formulation, our proposed DOTN outperforms previous
adversarial domain adaptation frameworks in a purely unsupervised manner.
Related papers
- Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - Activate and Reject: Towards Safe Domain Generalization under Category
Shift [71.95548187205736]
We study a practical problem of Domain Generalization under Category Shift (DGCS)
It aims to simultaneously detect unknown-class samples and classify known-class samples in the target domains.
Compared to prior DG works, we face two new challenges: 1) how to learn the concept of unknown'' during training with only source known-class samples, and 2) how to adapt the source-trained model to unseen environments.
arXiv Detail & Related papers (2023-10-07T07:53:12Z) - Taxonomy Adaptive Cross-Domain Adaptation in Medical Imaging via
Optimization Trajectory Distillation [73.83178465971552]
The success of automated medical image analysis depends on large-scale and expert-annotated training sets.
Unsupervised domain adaptation (UDA) has been raised as a promising approach to alleviate the burden of labeled data collection.
We propose optimization trajectory distillation, a unified approach to address the two technical challenges from a new perspective.
arXiv Detail & Related papers (2023-07-27T08:58:05Z) - Attentive Continuous Generative Self-training for Unsupervised Domain
Adaptive Medical Image Translation [12.080054869408213]
We develop a generative self-training framework for domain adaptive image translation with continuous value prediction and regression objectives.
We evaluate our framework on two cross-scanner/center, inter-subject translation tasks, including tagged-to-cine magnetic resonance (MR) image translation and T1-weighted MR-to-fractional anisotropy translation.
arXiv Detail & Related papers (2023-05-23T23:57:44Z) - CHATTY: Coupled Holistic Adversarial Transport Terms with Yield for
Unsupervised Domain Adaptation [1.87446486236017]
We propose a new technique called CHATTY: Coupled Holistic Adversarial Transport Terms with Yield for Unsupervised Domain Adaptation.
Adversarial training is commonly used for learning domain-invariant representations by reversing the gradients from a domain discriminator head to train the feature extractor layers of a neural network.
We introduce a sub-network which displaces the outputs of the source and target domain samples in a learnable manner.
arXiv Detail & Related papers (2023-04-19T13:00:23Z) - Cyclically Disentangled Feature Translation for Face Anti-spoofing [61.70377630461084]
We propose a novel domain adaptation method called cyclically disentangled feature translation network (CDFTN)
CDFTN generates pseudo-labeled samples that possess: 1) source domain-invariant liveness features and 2) target domain-specific content features, which are disentangled through domain adversarial training.
A robust classifier is trained based on the synthetic pseudo-labeled images under the supervision of source domain labels.
arXiv Detail & Related papers (2022-12-07T14:12:34Z) - Domain Adaptation with Adversarial Training on Penultimate Activations [82.9977759320565]
Enhancing model prediction confidence on unlabeled target data is an important objective in Unsupervised Domain Adaptation (UDA)
We show that this strategy is more efficient and better correlated with the objective of boosting prediction confidence than adversarial training on input images or intermediate features.
arXiv Detail & Related papers (2022-08-26T19:50:46Z) - Joint Attention-Driven Domain Fusion and Noise-Tolerant Learning for
Multi-Source Domain Adaptation [2.734665397040629]
Multi-source Unsupervised Domain Adaptation transfers knowledge from multiple source domains with labeled data to an unlabeled target domain.
The distribution discrepancy between different domains and the noisy pseudo-labels in the target domain both lead to performance bottlenecks.
We propose an approach that integrates Attention-driven Domain fusion and Noise-Tolerant learning (ADNT) to address the two issues mentioned above.
arXiv Detail & Related papers (2022-08-05T01:08:41Z) - Joint Distribution Alignment via Adversarial Learning for Domain
Adaptive Object Detection [11.262560426527818]
Unsupervised domain adaptive object detection aims to adapt a well-trained detector from its original source domain with rich labeled data to a new target domain with unlabeled data.
Recently, mainstream approaches perform this task through adversarial learning, yet still suffer from two limitations.
We propose a joint adaptive detection framework (JADF) to address the above challenges.
arXiv Detail & Related papers (2021-09-19T00:27:08Z) - Coarse to Fine: Domain Adaptive Crowd Counting via Adversarial Scoring
Network [58.05473757538834]
This paper proposes a novel adversarial scoring network (ASNet) to bridge the gap across domains from coarse to fine granularity.
Three sets of migration experiments show that the proposed methods achieve state-of-the-art counting performance.
arXiv Detail & Related papers (2021-07-27T14:47:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.