Adversarial Disentanglement of Speaker Representation for
Attribute-Driven Privacy Preservation
- URL: http://arxiv.org/abs/2012.04454v2
- Date: Fri, 23 Apr 2021 09:26:53 GMT
- Title: Adversarial Disentanglement of Speaker Representation for
Attribute-Driven Privacy Preservation
- Authors: Paul-Gauthier No\'e, Mohammad Mohammadamini, Driss Matrouf, Titouan
Parcollet, Andreas Nautsch, Jean-Fran\c{c}ois Bonastre
- Abstract summary: We introduce the concept of attribute-driven privacy preservation in speaker voice representation.
It allows a person to hide one or more personal aspects to a potential malicious interceptor and to the application provider.
We propose an adversarial autoencoding method that disentangles in the voice representation a given speaker attribute thus allowing its concealment.
- Score: 17.344080729609026
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In speech technologies, speaker's voice representation is used in many
applications such as speech recognition, voice conversion, speech synthesis
and, obviously, user authentication. Modern vocal representations of the
speaker are based on neural embeddings. In addition to the targeted
information, these representations usually contain sensitive information about
the speaker, like the age, sex, physical state, education level or ethnicity.
In order to allow the user to choose which information to protect, we introduce
in this paper the concept of attribute-driven privacy preservation in speaker
voice representation. It allows a person to hide one or more personal aspects
to a potential malicious interceptor and to the application provider. As a
first solution to this concept, we propose to use an adversarial autoencoding
method that disentangles in the voice representation a given speaker attribute
thus allowing its concealment. We focus here on the sex attribute for an
Automatic Speaker Verification (ASV) task. Experiments carried out using the
VoxCeleb datasets have shown that the proposed method enables the concealment
of this attribute while preserving ASV ability.
Related papers
- Are disentangled representations all you need to build speaker
anonymization systems? [0.0]
Speech signals contain a lot of sensitive information, such as the speaker's identity.
Speaker anonymization aims to transform a speech signal to remove the source speaker's identity while leaving the spoken content unchanged.
arXiv Detail & Related papers (2022-08-22T07:51:47Z) - Privacy-Preserving Speech Representation Learning using Vector
Quantization [0.0]
Speech signals contain a lot of sensitive information, such as the speaker's identity, which raises privacy concerns.
This paper aims to produce an anonymous representation while preserving speech recognition performance.
arXiv Detail & Related papers (2022-03-15T14:01:11Z) - Differentially Private Speaker Anonymization [44.90119821614047]
Sharing real-world speech utterances is key to the training and deployment of voice-based services.
Speaker anonymization aims to remove speaker information from a speech utterance while leaving its linguistic and prosodic attributes intact.
We show that disentanglement is indeed not perfect: linguistic and prosodic attributes still contain speaker information.
arXiv Detail & Related papers (2022-02-23T23:20:30Z) - An Attribute-Aligned Strategy for Learning Speech Representation [57.891727280493015]
We propose an attribute-aligned learning strategy to derive speech representation that can flexibly address these issues by attribute-selection mechanism.
Specifically, we propose a layered-representation variational autoencoder (LR-VAE), which factorizes speech representation into attribute-sensitive nodes.
Our proposed method achieves competitive performances on identity-free SER and a better performance on emotionless SV.
arXiv Detail & Related papers (2021-06-05T06:19:14Z) - Protecting gender and identity with disentangled speech representations [49.00162808063399]
We show that protecting gender information in speech is more effective than modelling speaker-identity information.
We present a novel way to encode gender information and disentangle two sensitive biometric identifiers.
arXiv Detail & Related papers (2021-04-22T13:31:41Z) - Voice Privacy with Smart Digital Assistants in Educational Settings [1.8369974607582578]
We design and evaluate a practical and efficient framework for voice privacy at the source.
The approach combines speaker identification (SID) and speech conversion methods to randomly disguise the identity of users right on the device that records the speech.
We evaluate the ASR performance of the conversion in terms of word error rate and show the promise of this framework in preserving the content of the input speech.
arXiv Detail & Related papers (2021-03-24T19:58:45Z) - Speaker De-identification System using Autoencoders and Adversarial
Training [58.720142291102135]
We propose a speaker de-identification system based on adversarial training and autoencoders.
Experimental results show that combining adversarial learning and autoencoders increase the equal error rate of a speaker verification system.
arXiv Detail & Related papers (2020-11-09T19:22:05Z) - Disentangled Speech Embeddings using Cross-modal Self-supervision [119.94362407747437]
We develop a self-supervised learning objective that exploits the natural cross-modal synchrony between faces and audio in video.
We construct a two-stream architecture which: (1) shares low-level features common to both representations; and (2) provides a natural mechanism for explicitly disentangling these factors.
arXiv Detail & Related papers (2020-02-20T14:13:12Z) - Speech Enhancement using Self-Adaptation and Multi-Head Self-Attention [70.82604384963679]
This paper investigates a self-adaptation method for speech enhancement using auxiliary speaker-aware features.
We extract a speaker representation used for adaptation directly from the test utterance.
arXiv Detail & Related papers (2020-02-14T05:05:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.