Do You Remember? Overcoming Catastrophic Forgetting for Fake Audio
Detection
- URL: http://arxiv.org/abs/2308.03300v1
- Date: Mon, 7 Aug 2023 05:05:49 GMT
- Title: Do You Remember? Overcoming Catastrophic Forgetting for Fake Audio
Detection
- Authors: Xiaohui Zhang, Jiangyan Yi, Jianhua Tao, Chenglong Wang, Chuyuan Zhang
- Abstract summary: We propose a continual learning algorithm for fake audio detection to overcome catastrophic forgetting.
When fine-tuning a detection network, our approach adaptively computes the direction of weight modification according to the ratio of genuine utterances and fake utterances.
Our method can easily be generalized to related fields, like speech emotion recognition.
- Score: 54.20974251478516
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Current fake audio detection algorithms have achieved promising performances
on most datasets. However, their performance may be significantly degraded when
dealing with audio of a different dataset. The orthogonal weight modification
to overcome catastrophic forgetting does not consider the similarity of genuine
audio across different datasets. To overcome this limitation, we propose a
continual learning algorithm for fake audio detection to overcome catastrophic
forgetting, called Regularized Adaptive Weight Modification (RAWM). When
fine-tuning a detection network, our approach adaptively computes the direction
of weight modification according to the ratio of genuine utterances and fake
utterances. The adaptive modification direction ensures the network can
effectively detect fake audio on the new dataset while preserving its knowledge
of old model, thus mitigating catastrophic forgetting. In addition, genuine
audio collected from quite different acoustic conditions may skew their feature
distribution, so we introduce a regularization constraint to force the network
to remember the old distribution in this regard. Our method can easily be
generalized to related fields, like speech emotion recognition. We also
evaluate our approach across multiple datasets and obtain a significant
performance improvement on cross-dataset experiments.
Related papers
- Statistics-aware Audio-visual Deepfake Detector [11.671275975119089]
Methods in audio-visualfake detection mostly assess the synchronization between audio and visual features.
We propose a statistical feature loss to enhance the discrimination capability of the model.
Experiments on the DFDC and FakeAVCeleb datasets demonstrate the relevance of the proposed method.
arXiv Detail & Related papers (2024-07-16T12:15:41Z) - Training-Free Deepfake Voice Recognition by Leveraging Large-Scale Pre-Trained Models [52.04189118767758]
Generalization is a main issue for current audio deepfake detectors.
In this paper we study the potential of large-scale pre-trained models for audio deepfake detection.
arXiv Detail & Related papers (2024-05-03T15:27:11Z) - What to Remember: Self-Adaptive Continual Learning for Audio Deepfake
Detection [53.063161380423715]
Existing detection models have shown remarkable success in discriminating known deepfake audio, but struggle when encountering new attack types.
We propose a continual learning approach called Radian Weight Modification (RWM) for audio deepfake detection.
arXiv Detail & Related papers (2023-12-15T09:52:17Z) - Adaptive Fake Audio Detection with Low-Rank Model Squeezing [50.7916414913962]
Traditional approaches, such as finetuning, are computationally intensive and pose a risk of impairing the acquired knowledge of known fake audio types.
We introduce the concept of training low-rank adaptation matrices tailored specifically to the newly emerging fake audio types.
Our approach offers several advantages, including reduced storage memory requirements and lower equal error rates.
arXiv Detail & Related papers (2023-06-08T06:06:42Z) - Improved DeepFake Detection Using Whisper Features [2.846767128062884]
We investigate the influence of Whisper automatic speech recognition model as a DF detection front-end.
We show that using Whisper-based features improves the detection for each model and outperforms recent results on the In-The-Wild dataset by reducing Equal Error Rate by 21%.
arXiv Detail & Related papers (2023-06-02T10:34:05Z) - Anomalous Sound Detection using Audio Representation with Machine ID
based Contrastive Learning Pretraining [52.191658157204856]
This paper uses contrastive learning to refine audio representations for each machine ID, rather than for each audio sample.
The proposed two-stage method uses contrastive learning to pretrain the audio representation model.
Experiments show that our method outperforms the state-of-the-art methods using contrastive learning or self-supervised classification.
arXiv Detail & Related papers (2023-04-07T11:08:31Z) - Automatic Change-Point Detection in Time Series via Deep Learning [8.43086628139493]
We show how to automatically generate new offline detection methods based on training a neural network.
We present theory that quantifies the error rate for such an approach, and how it depends on the amount of training data.
Our method also shows strong results in detecting and localising changes in activity based on accelerometer data.
arXiv Detail & Related papers (2022-11-07T20:59:14Z) - TASK3 DCASE2021 Challenge: Sound event localization and detection using
squeeze-excitation residual CNNs [4.4973334555746]
This study is based on the one carried out by the same team last year.
It has been decided to study how this technique improves each of the datasets.
This modification shows an improvement in the performance of the system compared to the baseline using MIC dataset.
arXiv Detail & Related papers (2021-07-30T11:34:15Z) - Continual Learning for Fake Audio Detection [62.54860236190694]
This paper proposes detecting fake without forgetting, a continual-learning-based method, to make the model learn new spoofing attacks incrementally.
Experiments are conducted on the ASVspoof 2019 dataset.
arXiv Detail & Related papers (2021-04-15T07:57:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.