Automated Label Generation for Time Series Classification with
Representation Learning: Reduction of Label Cost for Training
- URL: http://arxiv.org/abs/2107.05458v1
- Date: Mon, 12 Jul 2021 14:28:40 GMT
- Title: Automated Label Generation for Time Series Classification with
Representation Learning: Reduction of Label Cost for Training
- Authors: Soma Bandyopadhyay, Anish Datta, Arpan Pal
- Abstract summary: We propose a method to auto-generate labels of un-labelled time-series.
Our method is based on representation learning using Auto Encoded Compact Sequence.
It performs self-labelled in iterations, by learning latent structure, as well as synthetically boosting representative time-series.
- Score: 16.287885535569067
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Time-series generated by end-users, edge devices, and different wearables are
mostly unlabelled. We propose a method to auto-generate labels of un-labelled
time-series, exploiting very few representative labelled time-series. Our
method is based on representation learning using Auto Encoded Compact Sequence
(AECS) with a choice of best distance measure. It performs self-correction in
iterations, by learning latent structure, as well as synthetically boosting
representative time-series using Variational-Auto-Encoder (VAE) to improve the
quality of labels. We have experimented with UCR and UCI archives, public
real-world univariate, multivariate time-series taken from different
application domains. Experimental results demonstrate that the proposed method
is very close to the performance achieved by fully supervised classification.
The proposed method not only produces close to benchmark results but
outperforms the benchmark performance in some cases.
Related papers
- Pseudo-label Refinement for Improving Self-Supervised Learning Systems [22.276126184466207]
Self-supervised learning systems use clustering-based pseudo-labels to provide supervision without the need for human annotations.
The noise in these pseudo-labels caused by the clustering methods poses a challenge to the learning process leading to degraded performance.
We propose a pseudo-label refinement algorithm to address this issue.
arXiv Detail & Related papers (2024-10-18T07:47:59Z) - AEMLO: AutoEncoder-Guided Multi-Label Oversampling [6.255095509216069]
AEMLO is an AutoEncoder-guided Oversampling technique for imbalanced multi-label data.
We show that AEMLO outperforms the existing state-of-the-art methods with extensive empirical studies.
arXiv Detail & Related papers (2024-08-23T14:01:33Z) - AutoWS: Automated Weak Supervision Framework for Text Classification [1.748907524043535]
We propose a novel framework for increasing the efficiency of weak supervision process while decreasing the dependency on domain experts.
Our method requires a small set of labeled examples per label class and automatically creates a set of labeling functions to assign noisy labels to numerous unlabeled data.
arXiv Detail & Related papers (2023-02-07T07:12:05Z) - Gated recurrent units and temporal convolutional network for multilabel
classification [122.84638446560663]
This work proposes a new ensemble method for managing multilabel classification.
The core of the proposed approach combines a set of gated recurrent units and temporal convolutional neural networks trained with variants of the Adam gradients optimization approach.
arXiv Detail & Related papers (2021-10-09T00:00:16Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z) - Semi-Supervised Speech Recognition via Graph-based Temporal
Classification [59.58318952000571]
Semi-supervised learning has demonstrated promising results in automatic speech recognition by self-training.
The effectiveness of this approach largely relies on the pseudo-label accuracy.
Alternative ASR hypotheses of an N-best list can provide more accurate labels for an unlabeled speech utterance.
arXiv Detail & Related papers (2020-10-29T14:56:56Z) - Semi-supervised sequence classification through change point detection [22.14406516031776]
We propose a novel framework for semi-supervised learning in such contexts.
In an unsupervised manner, change point detection methods can be used to identify points within a sequence corresponding to likely class changes.
We show that change points provide examples of similar/dissimilar pairs of sequences which, when coupled with labeled, can be used in a semi-supervised classification setting.
arXiv Detail & Related papers (2020-09-24T17:23:13Z) - Change Point Detection in Time Series Data using Autoencoders with a
Time-Invariant Representation [69.34035527763916]
Change point detection (CPD) aims to locate abrupt property changes in time series data.
Recent CPD methods demonstrated the potential of using deep learning techniques, but often lack the ability to identify more subtle changes in the autocorrelation statistics of the signal.
We employ an autoencoder-based methodology with a novel loss function, through which the used autoencoders learn a partially time-invariant representation that is tailored for CPD.
arXiv Detail & Related papers (2020-08-21T15:03:21Z) - Unsupervised Person Re-identification via Multi-label Classification [55.65870468861157]
This paper formulates unsupervised person ReID as a multi-label classification task to progressively seek true labels.
Our method starts by assigning each person image with a single-class label, then evolves to multi-label classification by leveraging the updated ReID model for label prediction.
To boost the ReID model training efficiency in multi-label classification, we propose the memory-based multi-label classification loss (MMCL)
arXiv Detail & Related papers (2020-04-20T12:13:43Z) - Certified Robustness to Label-Flipping Attacks via Randomized Smoothing [105.91827623768724]
Machine learning algorithms are susceptible to data poisoning attacks.
We present a unifying view of randomized smoothing over arbitrary functions.
We propose a new strategy for building classifiers that are pointwise-certifiably robust to general data poisoning attacks.
arXiv Detail & Related papers (2020-02-07T21:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.