An Exponential Factorization Machine with Percentage Error Minimization
to Retail Sales Forecasting
- URL: http://arxiv.org/abs/2009.10619v1
- Date: Tue, 22 Sep 2020 15:21:38 GMT
- Title: An Exponential Factorization Machine with Percentage Error Minimization
to Retail Sales Forecasting
- Authors: Chongshou Li, Brenda Cheang, Zhixing Luo and Andrew Lim
- Abstract summary: This paper proposes a new approach to sales forecasting for new products with long lead time but short product life cycle.
An exponential factorization machine (EFM) sales forecast model is developed to solve this problem.
- Score: 3.7572675195649623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a new approach to sales forecasting for new products with
long lead time but short product life cycle. These SKUs are usually sold for
one season only, without any replenishments. An exponential factorization
machine (EFM) sales forecast model is developed to solve this problem which not
only considers SKU attributes, but also pairwise interactions. The EFM model is
significantly different from the original Factorization Machines (FM) from
two-fold: (1) the attribute-level formulation for explanatory variables and (2)
exponential formulation for the positive response variable. The attribute-level
formation excludes infeasible intra-attribute interactions and results in more
efficient feature engineering comparing with the conventional one-hot encoding,
while the exponential formulation is demonstrated more effective than the
log-transformation for the positive but not skewed distributed responses. In
order to estimate the parameters, percentage error squares (PES) and error
squares (ES) are minimized by a proposed adaptive batch gradient descent method
over the training set. Real-world data provided by a footwear retailer in
Singapore is used for testing the proposed approach. The forecasting
performance in terms of both mean absolute percentage error (MAPE) and mean
absolute error (MAE) compares favourably with not only off-the-shelf models but
also results reported by extant sales and demand forecasting studies. The
effectiveness of the proposed approach is also demonstrated by two external
public datasets. Moreover, we prove the theoretical relationships between PES
and ES minimization, and present an important property of the PES minimization
for regression models; that it trains models to underestimate data. This
property fits the situation of sales forecasting where unit-holding cost is
much greater than the unit-shortage cost.
Related papers
- Enhancing Retail Sales Forecasting with Optimized Machine Learning Models [0.0]
In retail sales forecasting, accurately predicting future sales is crucial for inventory management and strategic planning.
Recent advancements in machine learning (ML) provide more robust alternatives.
This research benefits from the power of ML, particularly Random Forest (RF), Gradient Boosting (GB), Support Vector Regression (SVR), and XGBoost.
arXiv Detail & Related papers (2024-10-17T17:11:33Z) - Distributionally Robust Optimization as a Scalable Framework to Characterize Extreme Value Distributions [22.765095010254118]
The goal of this paper is to develop distributionally robust optimization (DRO) estimators, specifically for multidimensional Extreme Value Theory (EVT) statistics.
In order to mitigate over-conservative estimates while enhancing out-of-sample performance, we study DRO estimators informed by semi-parametric max-stable constraints in the space of point processes.
Both approaches are validated using synthetically generated data, recovering prescribed characteristics, and verifying the efficacy of the proposed techniques.
arXiv Detail & Related papers (2024-07-31T19:45:27Z) - Low-rank finetuning for LLMs: A fairness perspective [54.13240282850982]
Low-rank approximation techniques have become the de facto standard for fine-tuning Large Language Models.
This paper investigates the effectiveness of these methods in capturing the shift of fine-tuning datasets from the initial pre-trained data distribution.
We show that low-rank fine-tuning inadvertently preserves undesirable biases and toxic behaviors.
arXiv Detail & Related papers (2024-05-28T20:43:53Z) - Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution [67.9215891673174]
We propose score entropy as a novel loss that naturally extends score matching to discrete spaces.
We test our Score Entropy Discrete Diffusion models on standard language modeling tasks.
arXiv Detail & Related papers (2023-10-25T17:59:12Z) - Self-Supervised Dataset Distillation for Transfer Learning [77.4714995131992]
We propose a novel problem of distilling an unlabeled dataset into a set of small synthetic samples for efficient self-supervised learning (SSL)
We first prove that a gradient of synthetic samples with respect to a SSL objective in naive bilevel optimization is textitbiased due to randomness originating from data augmentations or masking.
We empirically validate the effectiveness of our method on various applications involving transfer learning.
arXiv Detail & Related papers (2023-10-10T10:48:52Z) - Toward Fair Facial Expression Recognition with Improved Distribution
Alignment [19.442685015494316]
We present a novel approach to mitigate bias in facial expression recognition (FER) models.
Our method aims to reduce sensitive attribute information such as gender, age, or race, in the embeddings produced by FER models.
For the first time, we analyze the notion of attractiveness as an important sensitive attribute in FER models and demonstrate that FER models can indeed exhibit biases towards more attractive faces.
arXiv Detail & Related papers (2023-06-11T14:59:20Z) - Maintaining Stability and Plasticity for Predictive Churn Reduction [8.971668467496055]
We propose a solution called Accumulated Model Combination (AMC)
AMC is a general technique and we propose several instances of it, each having their own advantages depending on the model and data properties.
arXiv Detail & Related papers (2023-05-06T20:56:20Z) - Personalized Federated Learning under Mixture of Distributions [98.25444470990107]
We propose a novel approach to Personalized Federated Learning (PFL), which utilizes Gaussian mixture models (GMM) to fit the input data distributions across diverse clients.
FedGMM possesses an additional advantage of adapting to new clients with minimal overhead, and it also enables uncertainty quantification.
Empirical evaluations on synthetic and benchmark datasets demonstrate the superior performance of our method in both PFL classification and novel sample detection.
arXiv Detail & Related papers (2023-05-01T20:04:46Z) - Predicting with Confidence on Unseen Distributions [90.68414180153897]
We connect domain adaptation and predictive uncertainty literature to predict model accuracy on challenging unseen distributions.
We find that the difference of confidences (DoC) of a classifier's predictions successfully estimates the classifier's performance change over a variety of shifts.
We specifically investigate the distinction between synthetic and natural distribution shifts and observe that despite its simplicity DoC consistently outperforms other quantifications of distributional difference.
arXiv Detail & Related papers (2021-07-07T15:50:18Z) - A Locally Adaptive Interpretable Regression [7.4267694612331905]
Linear regression is one of the most interpretable prediction models.
In this work, we introduce a locally adaptive interpretable regression (LoAIR)
Our model achieves comparable or better predictive performance than the other state-of-the-art baselines.
arXiv Detail & Related papers (2020-05-07T09:26:14Z) - Mind the Trade-off: Debiasing NLU Models without Degrading the
In-distribution Performance [70.31427277842239]
We introduce a novel debiasing method called confidence regularization.
It discourages models from exploiting biases while enabling them to receive enough incentive to learn from all the training examples.
We evaluate our method on three NLU tasks and show that, in contrast to its predecessors, it improves the performance on out-of-distribution datasets.
arXiv Detail & Related papers (2020-05-01T11:22:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.