Quantifying Quality of Class-Conditional Generative Models in
Time-Series Domain
- URL: http://arxiv.org/abs/2210.07617v1
- Date: Fri, 14 Oct 2022 08:13:20 GMT
- Title: Quantifying Quality of Class-Conditional Generative Models in
Time-Series Domain
- Authors: Alireza Koochali, Maria Walch, Sankrutyayan Thota, Peter Schichtel,
Andreas Dengel, Sheraz Ahmed
- Abstract summary: We introduce the InceptionTime Score (ITS) and the Frechet InceptionTime Distance (FITD) to gauge the qualitative performance of class conditional generative models on the time-series domain.
We conduct extensive experiments on 80 different datasets to study the discriminative capabilities of proposed metrics.
- Score: 4.219228636765818
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative models are designed to address the data scarcity problem. Even
with the exploding amount of data, due to computational advancements, some
applications (e.g., health care, weather forecast, fault detection) still
suffer from data insufficiency, especially in the time-series domain. Thus
generative models are essential and powerful tools, but they still lack a
consensual approach for quality assessment. Such deficiency hinders the
confident application of modern implicit generative models on time-series data.
Inspired by assessment methods on the image domain, we introduce the
InceptionTime Score (ITS) and the Frechet InceptionTime Distance (FITD) to
gauge the qualitative performance of class conditional generative models on the
time-series domain. We conduct extensive experiments on 80 different datasets
to study the discriminative capabilities of proposed metrics alongside two
existing evaluation metrics: Train on Synthetic Test on Real (TSTR) and Train
on Real Test on Synthetic (TRTS). Extensive evaluation reveals that the
proposed assessment method, i.e., ITS and FITD in combination with TSTR, can
accurately assess class-conditional generative model performance.
Related papers
- Recurrent Neural Goodness-of-Fit Test for Time Series [8.22915954499148]
Time series data are crucial across diverse domains such as finance and healthcare.
Traditional evaluation metrics fall short due to the temporal dependencies and potential high dimensionality of the features.
We propose the REcurrent NeurAL (RENAL) Goodness-of-Fit test, a novel and statistically rigorous framework for evaluating generative time series models.
arXiv Detail & Related papers (2024-10-17T19:32:25Z) - How Hard is this Test Set? NLI Characterization by Exploiting Training Dynamics [49.9329723199239]
We propose a method for the automated creation of a challenging test set without relying on the manual construction of artificial and unrealistic examples.
We categorize the test set of popular NLI datasets into three difficulty levels by leveraging methods that exploit training dynamics.
When our characterization method is applied to the training set, models trained with only a fraction of the data achieve comparable performance to those trained on the full dataset.
arXiv Detail & Related papers (2024-10-04T13:39:21Z) - Synthesizing Multimodal Electronic Health Records via Predictive Diffusion Models [69.06149482021071]
We propose a novel EHR data generation model called EHRPD.
It is a diffusion-based model designed to predict the next visit based on the current one while also incorporating time interval estimation.
We conduct experiments on two public datasets and evaluate EHRPD from fidelity, privacy, and utility perspectives.
arXiv Detail & Related papers (2024-06-20T02:20:23Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - MADS: Modulated Auto-Decoding SIREN for time series imputation [9.673093148930874]
We propose MADS, a novel auto-decoding framework for time series imputation, built upon implicit neural representations.
We evaluate our model on two real-world datasets, and show that it outperforms state-of-the-art methods for time series imputation.
arXiv Detail & Related papers (2023-07-03T09:08:47Z) - From Static Benchmarks to Adaptive Testing: Psychometrics in AI Evaluation [60.14902811624433]
We discuss a paradigm shift from static evaluation methods to adaptive testing.
This involves estimating the characteristics and value of each test item in the benchmark and dynamically adjusting items in real-time.
We analyze the current approaches, advantages, and underlying reasons for adopting psychometrics in AI evaluation.
arXiv Detail & Related papers (2023-06-18T09:54:33Z) - Robustness and Generalization Performance of Deep Learning Models on
Cyber-Physical Systems: A Comparative Study [71.84852429039881]
Investigation focuses on the models' ability to handle a range of perturbations, such as sensor faults and noise.
We test the generalization and transfer learning capabilities of these models by exposing them to out-of-distribution (OOD) samples.
arXiv Detail & Related papers (2023-06-13T12:43:59Z) - Evaluating Model Performance in Medical Datasets Over Time [26.471486383140526]
This work proposes the Evaluation on Medical datasets Over Time (EMDOT) framework.
Inspired by the concept of backtesting, EMDOT simulates possible training procedures that practitioners might have been able to execute at each point in time.
We show how depending on the dataset, using all historical data may be ideal in many cases, whereas using a window of the most recent data could be advantageous in others.
arXiv Detail & Related papers (2023-05-22T19:16:00Z) - TimeVAE: A Variational Auto-Encoder for Multivariate Time Series
Generation [6.824692201913679]
We propose a novel architecture for synthetically generating time-series data with the use of Variversaational Auto-Encoders (VAEs)
The proposed architecture has several distinct properties: interpretability, ability to encode domain knowledge, and reduced training times.
arXiv Detail & Related papers (2021-11-15T21:42:14Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - Deep Semi-Supervised Learning for Time Series Classification [1.096924880299061]
We investigate the transferability of state-of-the-art deep semi-supervised models from image to time series classification.
We show that these transferred semi-supervised models show significant performance gains over strong supervised, semi-supervised and self-supervised alternatives.
arXiv Detail & Related papers (2021-02-06T17:40:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.