The Age of Correlated Features in Supervised Learning based Forecasting
- URL: http://arxiv.org/abs/2103.00092v1
- Date: Sat, 27 Feb 2021 00:10:49 GMT
- Title: The Age of Correlated Features in Supervised Learning based Forecasting
- Authors: MD Kamran Chowdhury Shisher, Heyang Qin, Lei Yang, Feng Yan, and Yin
Sun
- Abstract summary: In this paper, we analyze the impact of information freshness on supervised learning based forecasting.
In these applications, a neural network is trained to predict a time-varying target (e.g., solar power) based on multiple correlated features.
By using an information-theoretic approach, we prove that the minimum training loss is a function of the ages of the features.
- Score: 14.471447024994113
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: In this paper, we analyze the impact of information freshness on supervised
learning based forecasting. In these applications, a neural network is trained
to predict a time-varying target (e.g., solar power), based on multiple
correlated features (e.g., temperature, humidity, and cloud coverage). The
features are collected from different data sources and are subject to
heterogeneous and time-varying ages. By using an information-theoretic
approach, we prove that the minimum training loss is a function of the ages of
the features, where the function is not always monotonic. However, if the
empirical distribution of the training data is close to the distribution of a
Markov chain, then the training loss is approximately a non-decreasing age
function. Both the training loss and testing loss depict similar growth
patterns as the age increases. An experiment on solar power prediction is
conducted to validate our theory. Our theoretical and experimental results
suggest that it is beneficial to (i) combine the training data with different
age values into a large training dataset and jointly train the forecasting
decisions for these age values, and (ii) feed the age value as a part of the
input feature to the neural network.
Related papers
- Rethinking Benign Overfitting in Two-Layer Neural Networks [2.486161976966064]
We refine the feature-noise data model by incorporating class-dependent heterogeneous noise and re-examine the overfitting phenomenon in neural networks.
Our findings reveal that neural networks can leverage "data noise", previously deemed harmful, to learn implicit features that improve the classification accuracy for long-tailed data.
arXiv Detail & Related papers (2025-02-17T15:20:04Z) - When do Convolutional Neural Networks Stop Learning? [0.0]
Convolutional Neural Networks (CNNs) have demonstrated outstanding performance in computer vision tasks.
Current practice is to stop training when the training loss decreases and the gap between training and validation error increases.
This research work introduces a hypothesis that analyzes the data variation across all the layers of a CNN variant to anticipate its near-optimal learning capacity.
arXiv Detail & Related papers (2024-03-04T20:35:09Z) - The Mirrored Influence Hypothesis: Efficient Data Influence Estimation by Harnessing Forward Passes [30.30769701138665]
We introduce and explore the Mirrored Influence Hypothesis, highlighting a reciprocal nature of influence between training and test data.
Specifically, it suggests that evaluating the influence of training data on test predictions can be reformulated as an equivalent, yet inverse problem.
We introduce a new method for estimating the influence of training data, which requires calculating gradients for specific test samples, paired with a forward pass for each training point.
arXiv Detail & Related papers (2024-02-14T03:43:05Z) - On the Connection between Pre-training Data Diversity and Fine-tuning
Robustness [66.30369048726145]
We find that the primary factor influencing downstream effective robustness is data quantity.
We demonstrate our findings on pre-training distributions drawn from various natural and synthetic data sources.
arXiv Detail & Related papers (2023-07-24T05:36:19Z) - Stubborn Lexical Bias in Data and Models [50.79738900885665]
We use a new statistical method to examine whether spurious patterns in data appear in models trained on the data.
We apply an optimization approach to *reweight* the training data, reducing thousands of spurious correlations.
Surprisingly, though this method can successfully reduce lexical biases in the training data, we still find strong evidence of corresponding bias in the trained models.
arXiv Detail & Related papers (2023-06-03T20:12:27Z) - DCLP: Neural Architecture Predictor with Curriculum Contrastive Learning [5.2319020651074215]
We propose a Curricumum-guided Contrastive Learning framework for neural Predictor (DCLP)
Our method simplifies the contrastive task by designing a novel curriculum to enhance the stability of unlabeled training data distribution.
We experimentally demonstrate that DCLP has high accuracy and efficiency compared with existing predictors.
arXiv Detail & Related papers (2023-02-25T08:16:21Z) - Dataset Bias in Human Activity Recognition [57.91018542715725]
This contribution statistically curates the training data to assess to what degree the physical characteristics of humans influence HAR performance.
We evaluate the performance of a state-of-the-art convolutional neural network on two HAR datasets that vary in the sensors, activities, and recording for time-series HAR.
arXiv Detail & Related papers (2023-01-19T12:33:50Z) - Conformal prediction for the design problem [72.14982816083297]
In many real-world deployments of machine learning, we use a prediction algorithm to choose what data to test next.
In such settings, there is a distinct type of distribution shift between the training and test data.
We introduce a method to quantify predictive uncertainty in such settings.
arXiv Detail & Related papers (2022-02-08T02:59:12Z) - Convolutional generative adversarial imputation networks for
spatio-temporal missing data in storm surge simulations [86.5302150777089]
Generative Adversarial Imputation Nets (GANs) and GAN-based techniques have attracted attention as unsupervised machine learning methods.
We name our proposed method as Con Conval Generative Adversarial Imputation Nets (Conv-GAIN)
arXiv Detail & Related papers (2021-11-03T03:50:48Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Adversarial Regression Learning for Bone Age Estimation [6.942003070153651]
We propose an adversarial regression learning network (ARLNet) for bone age estimation.
Specifically, we first extract bone features from a fine-tuned Inception V3 neural network.
We then propose adversarial regression loss and feature reconstruction loss to guarantee the transition from training data to test data.
arXiv Detail & Related papers (2021-03-10T15:58:26Z) - Imbalanced Data Learning by Minority Class Augmentation using Capsule
Adversarial Networks [31.073558420480964]
We propose a method to restore the balance in imbalanced images, by coalescing two concurrent methods.
In our model, generative and discriminative networks play a novel competitive game.
The coalescing of capsule-GAN is effective at recognizing highly overlapping classes with much fewer parameters compared with the convolutional-GAN.
arXiv Detail & Related papers (2020-04-05T12:36:06Z) - Identifying and Compensating for Feature Deviation in Imbalanced Deep
Learning [59.65752299209042]
We investigate learning a ConvNet under such a scenario.
We found that a ConvNet significantly over-fits the minor classes.
We propose to incorporate class-dependent temperatures (CDT) training ConvNet.
arXiv Detail & Related papers (2020-01-06T03:52:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.