Demand Forecasting using Long Short-Term Memory Neural Networks
- URL: http://arxiv.org/abs/2008.08522v1
- Date: Wed, 19 Aug 2020 16:01:23 GMT
- Title: Demand Forecasting using Long Short-Term Memory Neural Networks
- Authors: Marta Go{\l}\k{a}bek, Robin Senge, and Rainer Neumann
- Abstract summary: Long short-term memory neural networks (LSTMs) are suitable for demand forecasting in the e-grocery retail sector.
Models showed better results for food products than the comparative models from both statistical and machine learning families.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we investigate to what extent long short-term memory neural
networks (LSTMs) are suitable for demand forecasting in the e-grocery retail
sector. For this purpose, univariate as well as multivariate LSTM-based models
were developed and tested for 100 fast-moving consumer goods in the context of
a master's thesis. On average, the developed models showed better results for
food products than the comparative models from both statistical and machine
learning families. Solely in the area of beverages random forest and linear
regression achieved slightly better results. This outcome suggests that LSTMs
can be used for demand forecasting at product level. The performance of the
models presented here goes beyond the current state of research, as can be seen
from the evaluations based on a data set that unfortunately has not been
publicly available to date.
Related papers
- Learning Augmentation Policies from A Model Zoo for Time Series Forecasting [58.66211334969299]
We introduce AutoTSAug, a learnable data augmentation method based on reinforcement learning.
By augmenting the marginal samples with a learnable policy, AutoTSAug substantially improves forecasting performance.
arXiv Detail & Related papers (2024-09-10T07:34:19Z) - F-FOMAML: GNN-Enhanced Meta-Learning for Peak Period Demand Forecasting with Proxy Data [65.6499834212641]
We formulate the demand prediction as a meta-learning problem and develop the Feature-based First-Order Model-Agnostic Meta-Learning (F-FOMAML) algorithm.
By considering domain similarities through task-specific metadata, our model improved generalization, where the excess risk decreases as the number of training tasks increases.
Compared to existing state-of-the-art models, our method demonstrates a notable improvement in demand prediction accuracy, reducing the Mean Absolute Error by 26.24% on an internal vending machine dataset and by 1.04% on the publicly accessible JD.com dataset.
arXiv Detail & Related papers (2024-06-23T21:28:50Z) - Few-Shot Load Forecasting Under Data Scarcity in Smart Grids: A Meta-Learning Approach [0.18641315013048293]
This paper proposes adapting an established model-agnostic meta-learning algorithm for short-term load forecasting.
The proposed method can rapidly adapt and generalize within any unknown load time series of arbitrary length.
The proposed model is evaluated using a dataset of historical load consumption data from real-world consumers.
arXiv Detail & Related papers (2024-06-09T18:59:08Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - Advanced Deep Regression Models for Forecasting Time Series Oil
Production [3.1383147856975633]
This research aims to develop advanced data-driven regression models using sequential convolutions and long short-term memory (LSTM) units.
It reveals that the LSTM-based sequence learning model can predict oil production better than the 1-D convolutional neural network (CNN) with mean absolute error (MAE) and R2 score of 111.16 and 0.98, respectively.
arXiv Detail & Related papers (2023-08-30T15:54:06Z) - Comparison of Transfer Learning based Additive Manufacturing Models via
A Case Study [3.759936323189418]
This paper defines a case study based on an open-source dataset about metal AM products.
Five TL methods are integrated with decision tree regression (DTR) and artificial neural network (ANN) to construct six TL-based models.
The comparisons are used to quantify the performance of applied TL methods and are discussed from the perspective of similarity, training data size, and data preprocessing.
arXiv Detail & Related papers (2023-05-17T00:29:25Z) - Robustness Analysis of Deep Learning Models for Population Synthesis [5.9106199000537645]
We present bootstrap confidence interval for the deep generative models to evaluate robustness to multiple datasets.
The models are implemented on multiple travel diaries of Montreal Origin- Destination Survey of 2008, 2013, and 2018.
Results show that the predictive errors of CTGAN have narrower confidence intervals indicating its robustness to multiple datasets.
arXiv Detail & Related papers (2022-11-23T22:55:55Z) - Comparing Prophet and Deep Learning to ARIMA in Forecasting Wholesale
Food Prices [0.0]
We study different techniques to forecast the sale prices of three food products applied by an Italian food wholesaler.
We consider ARIMA models and compare them to Prophet, a scalable forecasting tool developed by Facebook.
Our results indicate that ARIMA performs similarly to LSTM neural networks for the problem under study, while the combination of CNNs and LSTMs attains the best overall accuracy, but requires more time to be tuned.
arXiv Detail & Related papers (2021-07-23T15:13:31Z) - Back2Future: Leveraging Backfill Dynamics for Improving Real-time
Predictions in Future [73.03458424369657]
In real-time forecasting in public health, data collection is a non-trivial and demanding task.
'Backfill' phenomenon and its effect on model performance has been barely studied in the prior literature.
We formulate a novel problem and neural framework Back2Future that aims to refine a given model's predictions in real-time.
arXiv Detail & Related papers (2021-06-08T14:48:20Z) - Comparing Test Sets with Item Response Theory [53.755064720563]
We evaluate 29 datasets using predictions from 18 pretrained Transformer models on individual test examples.
We find that Quoref, HellaSwag, and MC-TACO are best suited for distinguishing among state-of-the-art models.
We also observe span selection task format, which is used for QA datasets like QAMR or SQuAD2.0, is effective in differentiating between strong and weak models.
arXiv Detail & Related papers (2021-06-01T22:33:53Z) - Learning representations with end-to-end models for improved remaining
useful life prognostics [64.80885001058572]
The remaining Useful Life (RUL) of equipment is defined as the duration between the current time and its failure.
We propose an end-to-end deep learning model based on multi-layer perceptron and long short-term memory layers (LSTM) to predict the RUL.
We will discuss how the proposed end-to-end model is able to achieve such good results and compare it to other deep learning and state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T16:45:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.