Efficient Training Data Generation for Phase-Based DOA Estimation
- URL: http://arxiv.org/abs/2011.04456v1
- Date: Mon, 9 Nov 2020 14:25:03 GMT
- Title: Efficient Training Data Generation for Phase-Based DOA Estimation
- Authors: Fabian H\"ubner, Wolfgang Mack, Emanu\"el A. P. Habets
- Abstract summary: Deep learning (DL) based direction of arrival (DOA) estimation is an active research topic and currently represents the state-of-the-art.
We propose a low complexity online data generation method to train DL models with a phase-based feature input.
By an evaluation using data from measured room impulse responses, we demonstrate that a model trained with the proposed training data generation method performs comparably to models trained with data generated based on the source-image method.
- Score: 8.035521056416243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning (DL) based direction of arrival (DOA) estimation is an active
research topic and currently represents the state-of-the-art. Usually, DL-based
DOA estimators are trained with recorded data or computationally expensive
generated data. Both data types require significant storage and excessive time
to, respectively, record or generate. We propose a low complexity online data
generation method to train DL models with a phase-based feature input. The data
generation method models the phases of the microphone signals in the frequency
domain by employing a deterministic model for the direct path and a statistical
model for the late reverberation of the room transfer function. By an
evaluation using data from measured room impulse responses, we demonstrate that
a model trained with the proposed training data generation method performs
comparably to models trained with data generated based on the source-image
method.
Related papers
- Synthesizing Multimodal Electronic Health Records via Predictive Diffusion Models [69.06149482021071]
We propose a novel EHR data generation model called EHRPD.
It is a diffusion-based model designed to predict the next visit based on the current one while also incorporating time interval estimation.
We conduct experiments on two public datasets and evaluate EHRPD from fidelity, privacy, and utility perspectives.
arXiv Detail & Related papers (2024-06-20T02:20:23Z) - Extracting Training Data from Unconditional Diffusion Models [76.85077961718875]
diffusion probabilistic models (DPMs) are being employed as mainstream models for generative artificial intelligence (AI)
We aim to establish a theoretical understanding of memorization in DPMs with 1) a memorization metric for theoretical analysis, 2) an analysis of conditional memorization with informative and random labels, and 3) two better evaluation metrics for measuring memorization.
Based on the theoretical analysis, we propose a novel data extraction method called textbfSurrogate condItional Data Extraction (SIDE) that leverages a trained on generated data as a surrogate condition to extract training data directly from unconditional diffusion models.
arXiv Detail & Related papers (2024-06-18T16:20:12Z) - Data Shapley in One Training Run [88.59484417202454]
Data Shapley provides a principled framework for attributing data's contribution within machine learning contexts.
Existing approaches require re-training models on different data subsets, which is computationally intensive.
This paper introduces In-Run Data Shapley, which addresses these limitations by offering scalable data attribution for a target model of interest.
arXiv Detail & Related papers (2024-06-16T17:09:24Z) - Distilled Datamodel with Reverse Gradient Matching [74.75248610868685]
We introduce an efficient framework for assessing data impact, comprising offline training and online evaluation stages.
Our proposed method achieves comparable model behavior evaluation while significantly speeding up the process compared to the direct retraining method.
arXiv Detail & Related papers (2024-04-22T09:16:14Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - Online Test-Time Adaptation of Spatial-Temporal Traffic Flow Forecasting [13.770733370640565]
This paper conducts the first study of the online test-time adaptation techniques for spatial-temporal traffic flow forecasting problems.
We propose an Adaptive Double Correction by Series Decomposition (ADCSD) method, which first decomposes the output of the trained model into seasonal and trend-cyclical parts.
In the proposed ADCSD method, instead of fine-tuning the whole trained model during the testing phase, a lite network is attached after the trained model, and only the lite network is fine-tuned in the testing process each time a data entry is observed.
arXiv Detail & Related papers (2024-01-08T12:04:39Z) - Diffusion-Model-Assisted Supervised Learning of Generative Models for
Density Estimation [10.793646707711442]
We present a framework for training generative models for density estimation.
We use the score-based diffusion model to generate labeled data.
Once the labeled data are generated, we can train a simple fully connected neural network to learn the generative model in the supervised manner.
arXiv Detail & Related papers (2023-10-22T23:56:19Z) - Representation Transfer Learning via Multiple Pre-trained models for
Linear Regression [3.5788754401889014]
We consider the problem of learning a linear regression model on a data domain of interest (target) given few samples.
To aid learning, we are provided with a set of pre-trained regression models that are trained on potentially different data domains.
We propose a representation transfer based learning method for constructing the target model.
arXiv Detail & Related papers (2023-05-25T19:35:24Z) - Graph-Based Model-Agnostic Data Subsampling for Recommendation Systems [29.713557081485995]
Data subsampling is widely used to speed up the training of recommendation systems.
Most subsampling methods are model-based and often require a pre-trained pilot model to measure data importance.
We propose model-agnostic data subsampling methods by only exploring input data structure represented by graphs.
arXiv Detail & Related papers (2023-05-25T18:00:15Z) - Transfer Learning with Uncertainty Quantification: Random Effect
Calibration of Source to Target (RECaST) [1.8047694351309207]
We develop a statistical framework for model predictions based on transfer learning, called RECaST.
We mathematically and empirically demonstrate the validity of our RECaST approach for transfer learning between linear models.
We examine our method's performance in a simulation study and in an application to real hospital data.
arXiv Detail & Related papers (2022-11-29T19:39:47Z) - Learning from aggregated data with a maximum entropy model [73.63512438583375]
We show how a new model, similar to a logistic regression, may be learned from aggregated data only by approximating the unobserved feature distribution with a maximum entropy hypothesis.
We present empirical evidence on several public datasets that the model learned this way can achieve performances comparable to those of a logistic model trained with the full unaggregated data.
arXiv Detail & Related papers (2022-10-05T09:17:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.