Multi-task Meta Label Correction for Time Series Prediction
- URL: http://arxiv.org/abs/2303.08103v3
- Date: Sun, 18 Feb 2024 07:48:23 GMT
- Title: Multi-task Meta Label Correction for Time Series Prediction
- Authors: Luxuan Yang, Ting Gao, Wei Wei, Min Dai, Cheng Fang, Jinqiao Duan
- Abstract summary: We create a label correction method to time series data with meta-learning under a multi-task framework.
Results show that our method is more effective and accurate than some existing label correction techniques.
- Score: 10.08574256346388
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Time series classification faces two unavoidable problems. One is partial
feature information and the other is poor label quality, which may affect model
performance. To address the above issues, we create a label correction method
to time series data with meta-learning under a multi-task framework. There are
three main contributions. First, we train the label correction model with a
two-branch neural network in the outer loop. While in the model-agnostic inner
loop, we use pre-existing classification models in a multi-task way and jointly
update the meta-knowledge so as to help us achieve adaptive labeling on complex
time series. Second, we devise new data visualization methods for both image
patterns of the historical data and data in the prediction horizon. Finally, we
test our method with various financial datasets, including XOM, S\&P500, and
SZ50. Results show that our method is more effective and accurate than some
existing label correction techniques.
Related papers
- Label-Guided Imputation via Forest-Based Proximities for Improved Time Series Classification [1.6863755729554886]
We provide a framework for missing data imputation in the context of time series classification.<n>We define a means of imputing missing values conditional upon labels.<n>We show that imputation using this method generally provides richer information leading to higher classification accuracies.
arXiv Detail & Related papers (2025-09-26T20:47:00Z) - Towards Micro-Action Recognition with Limited Annotations: An Asynchronous Pseudo Labeling and Training Approach [35.32024173141412]
We introduce the setting of Semi-Supervised MAR (SSMAR), where only a part of samples are labeled.
Traditional Semi-Supervised Learning (SSL) methods tend to overfit on inaccurate pseudo-labels, leading to error accumulation and degraded performance.
We propose Asynchronous Pseudo Labeling and Training (APLT), which explicitly separates the pseudo-labeling process from model training.
arXiv Detail & Related papers (2025-04-10T14:22:15Z) - Efficient Grammatical Error Correction Via Multi-Task Training and
Optimized Training Schedule [55.08778142798106]
We propose auxiliary tasks that exploit the alignment between the original and corrected sentences.
We formulate each task as a sequence-to-sequence problem and perform multi-task training.
We find that the order of datasets used for training and even individual instances within a dataset may have important effects on the final performance.
arXiv Detail & Related papers (2023-11-20T14:50:12Z) - Improving Zero-Shot Models with Label Distribution Priors [33.51714665243138]
We propose a new approach, CLIPPR, which adapts zero-shot models for regression and classification on unlabelled datasets.
We demonstrate an improvement of 28% in mean absolute error on the UTK age regression task.
We also present promising results for classification benchmarks, improving the classification accuracy on the ImageNet dataset by 2.83%, without using any labels.
arXiv Detail & Related papers (2022-12-01T18:59:03Z) - SEAL : Interactive Tool for Systematic Error Analysis and Labeling [26.803598323167382]
This paper introduces an interactive Systematic Error Analysis and Labeling (seal) tool.
It uses a two-step approach to first identify high error slices of data and then, in the second step, introduce methods to give human-understandable semantics to those underperforming slices.
We explore a variety of methods for coming up with coherent semantics for the error groups using language models for semantic labeling and a text-to-image model for generating visual features.
arXiv Detail & Related papers (2022-10-11T23:51:44Z) - Active Learning by Feature Mixing [52.16150629234465]
We propose a novel method for batch active learning called ALFA-Mix.
We identify unlabelled instances with sufficiently-distinct features by seeking inconsistencies in predictions.
We show that inconsistencies in these predictions help discovering features that the model is unable to recognise in the unlabelled instances.
arXiv Detail & Related papers (2022-03-14T12:20:54Z) - DapStep: Deep Assignee Prediction for Stack Trace Error rePresentation [61.99379022383108]
We propose new deep learning models to solve the bug triage problem.
The models are based on a bidirectional recurrent neural network with attention and on a convolutional neural network.
To improve the quality of ranking, we propose using additional information from version control system annotations.
arXiv Detail & Related papers (2022-01-14T00:16:57Z) - Cross-Model Pseudo-Labeling for Semi-Supervised Action Recognition [98.25592165484737]
We propose a more effective pseudo-labeling scheme, called Cross-Model Pseudo-Labeling (CMPL)
CMPL achieves $17.6%$ and $25.1%$ Top-1 accuracy on Kinetics-400 and UCF-101 using only the RGB modality and $1%$ labeled data, respectively.
arXiv Detail & Related papers (2021-12-17T18:59:41Z) - Towards Good Practices for Efficiently Annotating Large-Scale Image
Classification Datasets [90.61266099147053]
We investigate efficient annotation strategies for collecting multi-class classification labels for a large collection of images.
We propose modifications and best practices aimed at minimizing human labeling effort.
Simulated experiments on a 125k image subset of the ImageNet100 show that it can be annotated to 80% top-1 accuracy with 0.35 annotations per image on average.
arXiv Detail & Related papers (2021-04-26T16:29:32Z) - How to distribute data across tasks for meta-learning? [59.608652082495624]
We show that the optimal number of data points per task depends on the budget, but it converges to a unique constant value for large budgets.
Our results suggest a simple and efficient procedure for data collection.
arXiv Detail & Related papers (2021-03-15T15:38:47Z) - Meta-Regularization by Enforcing Mutual-Exclusiveness [0.8057006406834467]
We propose a regularization technique for meta-learning models that gives the model designer more control over the information flow during meta-training.
Our proposed regularization function shows an accuracy boost of $sim$ $36%$ on the Omniglot dataset.
arXiv Detail & Related papers (2021-01-24T22:57:19Z) - Concept Drift and Covariate Shift Detection Ensemble with Lagged Labels [22.826118321715455]
It is important to detect changes and retrain the model in time.
Current methods have three weaknesses: 1) using only classification error rate as signal, 2) assuming ground truth labels are immediately available after features from samples are received and 3) unable to decide what data to use to retrain the model when change occurs.
We address the first problem by utilizing six different signals to capture a wide range of characteristics of data, and we address the second problem by allowing lag of labels, where labels of corresponding features are received after a lag in time.
For the third problem, our proposed method automatically decides what data to use to retrain
arXiv Detail & Related papers (2020-12-08T21:57:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.