Empirical Evaluation of Data Augmentations for Biobehavioral Time Series
Data with Deep Learning
- URL: http://arxiv.org/abs/2210.06701v1
- Date: Thu, 13 Oct 2022 03:40:12 GMT
- Title: Empirical Evaluation of Data Augmentations for Biobehavioral Time Series
Data with Deep Learning
- Authors: Huiyuan Yang, Han Yu, Akane Sano
- Abstract summary: Data augmentation (DA) is a critical step for the success of deep learning models on biobehavioral time series data.
We first systematically review eight basic DA methods for biobehavioral time series data, and evaluate the effects on seven datasets with three backbones.
Next, we explore adapting more recent DA techniques to biobehavioral time series data by designing a new policy architecture.
- Score: 16.84326709739788
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has performed remarkably well on many tasks recently. However,
the superior performance of deep models relies heavily on the availability of a
large number of training data, which limits the wide adaptation of deep models
on various clinical and affective computing tasks, as the labeled data are
usually very limited. As an effective technique to increase the data
variability and thus train deep models with better generalization, data
augmentation (DA) is a critical step for the success of deep learning models on
biobehavioral time series data. However, the effectiveness of various DAs for
different datasets with different tasks and deep models is understudied for
biobehavioral time series data. In this paper, we first systematically review
eight basic DA methods for biobehavioral time series data, and evaluate the
effects on seven datasets with three backbones. Next, we explore adapting more
recent DA techniques (i.e., automatic augmentation, random augmentation) to
biobehavioral time series data by designing a new policy architecture
applicable to time series data. Last, we try to answer the question of why a DA
is effective (or not) by first summarizing two desired attributes for
augmentations (challenging and faithful), and then utilizing two metrics to
quantitatively measure the corresponding attributes, which can guide us in the
search for more effective DA for biobehavioral time series data by designing
more challenging but still faithful transformations. Our code and results are
available at Link.
Related papers
- Deep Time Series Models: A Comprehensive Survey and Benchmark [74.28364194333447]
Time series data is of great significance in real-world scenarios.
Recent years have witnessed remarkable breakthroughs in the time series community.
We release Time Series Library (TSLib) as a fair benchmark of deep time series models for diverse analysis tasks.
arXiv Detail & Related papers (2024-07-18T08:31:55Z) - Data Augmentation for Multivariate Time Series Classification: An Experimental Study [1.5390962520179197]
Despite the limited size of these datasets, we achieved classification accuracy improvements in 10 out of 13 datasets using the Rocket and InceptionTime models.
This highlights the essential role of sufficient data in training effective models, paralleling the advancements seen in computer vision.
arXiv Detail & Related papers (2024-06-10T17:58:02Z) - LESS: Selecting Influential Data for Targeted Instruction Tuning [64.78894228923619]
We propose LESS, an efficient algorithm to estimate data influences and perform Low-rank gradiEnt Similarity Search for instruction data selection.
We show that training on a LESS-selected 5% of the data can often outperform training on the full dataset across diverse downstream tasks.
Our method goes beyond surface form cues to identify data that the necessary reasoning skills for the intended downstream application.
arXiv Detail & Related papers (2024-02-06T19:18:04Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - MADS: Modulated Auto-Decoding SIREN for time series imputation [9.673093148930874]
We propose MADS, a novel auto-decoding framework for time series imputation, built upon implicit neural representations.
We evaluate our model on two real-world datasets, and show that it outperforms state-of-the-art methods for time series imputation.
arXiv Detail & Related papers (2023-07-03T09:08:47Z) - Time Series Contrastive Learning with Information-Aware Augmentations [57.45139904366001]
A key component of contrastive learning is to select appropriate augmentations imposing some priors to construct feasible positive samples.
How to find the desired augmentations of time series data that are meaningful for given contrastive learning tasks and datasets remains an open question.
We propose a new contrastive learning approach with information-aware augmentations, InfoTS, that adaptively selects optimal augmentations for time series representation learning.
arXiv Detail & Related papers (2023-03-21T15:02:50Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - Data Augmentation techniques in time series domain: A survey and
taxonomy [0.20971479389679332]
Deep neural networks used to work with time series heavily depend on the size and consistency of the datasets used in training.
This work systematically reviews the current state-of-the-art in the area to provide an overview of all available algorithms.
The ultimate aim of this study is to provide a summary of the evolution and performance of areas that produce better results to guide future researchers in this field.
arXiv Detail & Related papers (2022-06-25T17:09:00Z) - Diverse Complexity Measures for Dataset Curation in Self-driving [80.55417232642124]
We propose a new data selection method that exploits a diverse set of criteria that quantize interestingness of traffic scenes.
Our experiments show that the proposed curation pipeline is able to select datasets that lead to better generalization and higher performance.
arXiv Detail & Related papers (2021-01-16T23:45:02Z) - Deep learning for time series classification [2.0305676256390934]
Time series analysis allows us to visualize and understand the evolution of a process over time.
Time series classification consists of constructing algorithms dedicated to automatically label time series data.
Deep learning has emerged as one of the most effective methods for tackling the supervised classification task.
arXiv Detail & Related papers (2020-10-01T17:38:40Z) - Time Series Data Augmentation for Deep Learning: A Survey [35.2161833151567]
We systematically review different data augmentation methods for time series data.
We empirically compare different data augmentation methods for different tasks including time series classification, anomaly detection, and forecasting.
arXiv Detail & Related papers (2020-02-27T23:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.