How Does Data Freshness Affect Real-time Supervised Learning?
- URL: http://arxiv.org/abs/2208.06948v1
- Date: Mon, 15 Aug 2022 00:14:13 GMT
- Title: How Does Data Freshness Affect Real-time Supervised Learning?
- Authors: Md Kamran Chowdhury Shisher and Yin Sun
- Abstract summary: We show that the performance of real-time supervised learning degrades monotonically as the feature becomes stale.
To minimize the inference error in real-time, we propose a new "selection-from-buffer" model for sending the features.
Data-driven evaluations are presented to illustrate the benefits of the proposed scheduling algorithms.
- Score: 15.950108699395077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we analyze the impact of data freshness on real-time
supervised learning, where a neural network is trained to infer a time-varying
target (e.g., the position of the vehicle in front) based on features (e.g.,
video frames) observed at a sensing node (e.g., camera or lidar). One might
expect that the performance of real-time supervised learning degrades
monotonically as the feature becomes stale. Using an information-theoretic
analysis, we show that this is true if the feature and target data sequence can
be closely approximated as a Markov chain; it is not true if the data sequence
is far from Markovian. Hence, the prediction error of real-time supervised
learning is a function of the Age of Information (AoI), where the function
could be non-monotonic. Several experiments are conducted to illustrate the
monotonic and non-monotonic behaviors of the prediction error. To minimize the
inference error in real-time, we propose a new "selection-from-buffer" model
for sending the features, which is more general than the "generate-at-will"
model used in earlier studies. By using Gittins and Whittle indices,
low-complexity scheduling strategies are developed to minimize the inference
error, where a new connection between the Gittins index theory and Age of
Information (AoI) minimization is discovered. These scheduling results hold (i)
for minimizing general AoI functions (monotonic or non-monotonic) and (ii) for
general feature transmission time distributions. Data-driven evaluations are
presented to illustrate the benefits of the proposed scheduling algorithms.
Related papers
- Multiple Descents in Unsupervised Learning: The Role of Noise, Domain Shift and Anomalies [14.399035468023161]
We study the presence of double descent in unsupervised learning, an area that has received little attention and is not yet fully understood.
We use synthetic and real data and identify model-wise, epoch-wise, and sample-wise double descent for various applications.
arXiv Detail & Related papers (2024-06-17T16:24:23Z) - Timely Communications for Remote Inference [16.671201899392585]
We analyze the impact of data freshness on remote inference systems.
We propose a new "selection-from-buffer" model for sending the features.
We also design low-complexity scheduling policies to improve inference performance.
arXiv Detail & Related papers (2024-04-25T01:53:21Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Uncovering the Missing Pattern: Unified Framework Towards Trajectory
Imputation and Prediction [60.60223171143206]
Trajectory prediction is a crucial undertaking in understanding entity movement or human behavior from observed sequences.
Current methods often assume that the observed sequences are complete while ignoring the potential for missing values.
This paper presents a unified framework, the Graph-based Conditional Variational Recurrent Neural Network (GC-VRNN), which can perform trajectory imputation and prediction simultaneously.
arXiv Detail & Related papers (2023-03-28T14:27:27Z) - Probing Representation Forgetting in Supervised and Unsupervised
Continual Learning [14.462797749666992]
Catastrophic forgetting is associated with an abrupt loss of knowledge previously learned by a model.
We show that representation forgetting can lead to new insights on the effect of model capacity and loss function used in continual learning.
arXiv Detail & Related papers (2022-03-24T23:06:08Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Networked Time Series Prediction with Incomplete Data [59.45358694862176]
We propose NETS-ImpGAN, a novel deep learning framework that can be trained on incomplete data with missing values in both history and future.
We conduct extensive experiments on three real-world datasets under different missing patterns and missing rates.
arXiv Detail & Related papers (2021-10-05T18:20:42Z) - Imputation-Free Learning from Incomplete Observations [73.15386629370111]
We introduce the importance of guided gradient descent (IGSGD) method to train inference from inputs containing missing values without imputation.
We employ reinforcement learning (RL) to adjust the gradients used to train the models via back-propagation.
Our imputation-free predictions outperform the traditional two-step imputation-based predictions using state-of-the-art imputation methods.
arXiv Detail & Related papers (2021-07-05T12:44:39Z) - Learning Transition Models with Time-delayed Causal Relations [17.494609199646813]
The proposed algorithm initially predicts observations with the Markov assumption.
hidden variables are memory units that keep track of pertinent past events.
Experiments on simulated and real robotic tasks show that this method significantly improves over current techniques.
arXiv Detail & Related papers (2020-08-04T14:35:11Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z) - Conditional Mutual information-based Contrastive Loss for Financial Time
Series Forecasting [12.0855096102517]
We present a representation learning framework for financial time series forecasting.
In this paper, we propose to first learn compact representations from time series data, then use the learned representations to train a simpler model for predicting time series movements.
arXiv Detail & Related papers (2020-02-18T15:24:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.