Unsupervised Cross-Lingual Speech Emotion Recognition Using
DomainAdversarial Neural Network
- URL: http://arxiv.org/abs/2012.11174v1
- Date: Mon, 21 Dec 2020 08:21:11 GMT
- Title: Unsupervised Cross-Lingual Speech Emotion Recognition Using
DomainAdversarial Neural Network
- Authors: Xiong Cai, Zhiyong Wu, Kuo Zhong, Bin Su, Dongyang Dai, Helen Meng
- Abstract summary: Cross-domain Speech Emotion Recog-nition (SER) is still a challenging taskdue to the distribution shift between source and target domains.
We propose a Domain Adversarial Neural Net-work (DANN) based approach to mitigate this distribution shiftproblem for cross-lingual SER.
- Score: 48.1535353007371
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: By using deep learning approaches, Speech Emotion Recog-nition (SER) on a
single domain has achieved many excellentresults. However, cross-domain SER is
still a challenging taskdue to the distribution shift between source and target
domains.In this work, we propose a Domain Adversarial Neural Net-work (DANN)
based approach to mitigate this distribution shiftproblem for cross-lingual
SER. Specifically, we add a languageclassifier and gradient reversal layer
after the feature extractor toforce the learned representation both
language-independent andemotion-meaningful. Our method is unsupervised, i. e.,
labelson target language are not required, which makes it easier to ap-ply our
method to other languages. Experimental results showthe proposed method
provides an average absolute improve-ment of 3.91% over the baseline system for
arousal and valenceclassification task. Furthermore, we find that batch
normaliza-tion is beneficial to the performance gain of DANN. Thereforewe also
explore the effect of different ways of data combinationfor batch
normalization.
Related papers
- Reducing Source-Private Bias in Extreme Universal Domain Adaptation [11.875619863954238]
Universal Domain Adaptation (UniDA) aims to transfer knowledge from a labeled source domain to an unlabeled target domain.
We show that state-of-the-art methods struggle when the source domain has significantly more non-overlapping classes than overlapping ones.
We propose using self-supervised learning to preserve the structure of the target data.
arXiv Detail & Related papers (2024-10-15T04:51:37Z) - FIT: Frequency-based Image Translation for Domain Adaptive Object
Detection [8.635264598464355]
We propose a novel Frequency-based Image Translation (FIT) framework for Domain adaptive object detection (DAOD)
First, by keeping domain-invariant frequency components and swapping domain-specific ones, we conduct image translation to reduce domain shift at the input level.
Second, hierarchical adversarial feature learning is utilized to further mitigate the domain gap at the feature level.
arXiv Detail & Related papers (2023-03-07T07:30:08Z) - Self-training through Classifier Disagreement for Cross-Domain Opinion
Target Extraction [62.41511766918932]
Opinion target extraction (OTE) or aspect extraction (AE) is a fundamental task in opinion mining.
Recent work focus on cross-domain OTE, which is typically encountered in real-world scenarios.
We propose a new SSL approach that opts for selecting target samples whose model output from a domain-specific teacher and student network disagrees on the unlabelled target data.
arXiv Detail & Related papers (2023-02-28T16:31:17Z) - Semi-supervised cross-lingual speech emotion recognition [26.544999411050036]
Cross-lingual Speech Emotion Recognition remains a challenge in real-world applications.
We propose a Semi-Supervised Learning (SSL) method for cross-lingual emotion recognition when only few labeled examples in the target domain are available.
Our method is based on a Transformer and it adapts to the new domain by exploiting a pseudo-labeling strategy on the unlabeled utterances.
arXiv Detail & Related papers (2022-07-14T09:24:55Z) - Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training
for Road Segmentation of Remote Sensing Images [93.50240389540252]
Road segmentation from remote sensing images is a challenging task with wide ranges of application potentials.
We propose a novel stagewise domain adaptation model called RoadDA to address the domain shift (DS) issue in this field.
Experiment results on two benchmarks demonstrate that RoadDA can efficiently reduce the domain gap and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2021-08-28T09:29:14Z) - OVANet: One-vs-All Network for Universal Domain Adaptation [78.86047802107025]
Existing methods manually set a threshold to reject unknown samples based on validation or a pre-defined ratio of unknown samples.
We propose a method to learn the threshold using source samples and to adapt it to the target domain.
Our idea is that a minimum inter-class distance in the source domain should be a good threshold to decide between known or unknown in the target.
arXiv Detail & Related papers (2021-04-07T18:36:31Z) - Effective Label Propagation for Discriminative Semi-Supervised Domain
Adaptation [76.41664929948607]
Semi-supervised domain adaptation (SSDA) methods have demonstrated great potential in large-scale image classification tasks.
We present a novel and effective method to tackle this problem by using effective inter-domain and intra-domain semantic information propagation.
Our source code and pre-trained models will be released soon.
arXiv Detail & Related papers (2020-12-04T14:28:19Z) - Alleviating Semantic-level Shift: A Semi-supervised Domain Adaptation
Method for Semantic Segmentation [97.8552697905657]
A key challenge of this task is how to alleviate the data distribution discrepancy between the source and target domains.
We propose Alleviating Semantic-level Shift (ASS), which can successfully promote the distribution consistency from both global and local views.
We apply our ASS to two domain adaptation tasks, from GTA5 to Cityscapes and from Synthia to Cityscapes.
arXiv Detail & Related papers (2020-04-02T03:25:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.