Data Augmentation Strategies for Improving Sequential Recommender
Systems
- URL: http://arxiv.org/abs/2203.14037v1
- Date: Sat, 26 Mar 2022 09:58:14 GMT
- Title: Data Augmentation Strategies for Improving Sequential Recommender
Systems
- Authors: Joo-yeong Song, Bongwon Suh
- Abstract summary: Sequential recommender systems have recently achieved significant performance improvements with the exploitation of deep learning (DL) based methods.
We propose a set of data augmentation strategies, all of which transform original item sequences in the way of direct corruption.
Experiments on the latest DL-based model show that applying data augmentation can help the model generalize better.
- Score: 7.986899327513767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequential recommender systems have recently achieved significant performance
improvements with the exploitation of deep learning (DL) based methods.
However, although various DL-based methods have been introduced, most of them
only focus on the transformations of network structure, neglecting the
importance of other influential factors including data augmentation. Obviously,
DL-based models require a large amount of training data in order to estimate
parameters well and achieve high performances, which leads to the early efforts
to increase the training data through data augmentation in computer vision and
speech domains. In this paper, we seek to figure out that various data
augmentation strategies can improve the performance of sequential recommender
systems, especially when the training dataset is not large enough. To this end,
we propose a simple set of data augmentation strategies, all of which transform
original item sequences in the way of direct corruption and describe how data
augmentation changes the performance. Extensive experiments on the latest
DL-based model show that applying data augmentation can help the model
generalize better, and it can be significantly effective to boost model
performances especially when the amount of training data is small. Furthermore,
it is shown that our proposed strategies can improve performances to a better
or competitive level to existing strategies suggested in the prior works.
Related papers
- Data Efficiency for Large Recommendation Models [4.799343040337817]
Large recommendation models (LRMs) are fundamental to the multi-billion dollar online advertising industry.
The massive scale of data directly impacts both computational costs and the speed at which new methods can be evaluated.
This paper presents actionable principles and high-level frameworks to guide practitioners in optimizing training data requirements.
arXiv Detail & Related papers (2024-10-08T22:18:00Z) - Data Augmentation for Multivariate Time Series Classification: An Experimental Study [1.5390962520179197]
Despite the limited size of these datasets, we achieved classification accuracy improvements in 10 out of 13 datasets using the Rocket and InceptionTime models.
This highlights the essential role of sufficient data in training effective models, paralleling the advancements seen in computer vision.
arXiv Detail & Related papers (2024-06-10T17:58:02Z) - Rethinking Overlooked Aspects in Vision-Language Models [32.525916879333145]
Recent advancements in vision-language models (LVLMs) have been substantial.
Recent works mainly focus on introducing more pre-training and instruction tuning data to improve model's performance.
This paper delves into the often-neglected aspects of data efficiency during pre-training and the selection process for instruction tuning datasets.
arXiv Detail & Related papers (2024-05-20T07:53:41Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - Small Dataset, Big Gains: Enhancing Reinforcement Learning by Offline
Pre-Training with Model Based Augmentation [59.899714450049494]
offline pre-training can produce sub-optimal policies and lead to degraded online reinforcement learning performance.
We propose a model-based data augmentation strategy to maximize the benefits of offline reinforcement learning pre-training and reduce the scale of data needed to be effective.
arXiv Detail & Related papers (2023-12-15T14:49:41Z) - Data-Centric Long-Tailed Image Recognition [49.90107582624604]
Long-tail models exhibit a strong demand for high-quality data.
Data-centric approaches aim to enhance both the quantity and quality of data to improve model performance.
There is currently a lack of research into the underlying mechanisms explaining the effectiveness of information augmentation.
arXiv Detail & Related papers (2023-11-03T06:34:37Z) - Phased Data Augmentation for Training a Likelihood-Based Generative Model with Limited Data [0.0]
Generative models excel in creating realistic images, yet their dependency on extensive datasets for training presents significant challenges.
Current data-efficient methods largely focus on GAN architectures, leaving a gap in training other types of generative models.
"phased data augmentation" is a novel technique that addresses this gap by optimizing training in limited data scenarios without altering the inherent data distribution.
arXiv Detail & Related papers (2023-05-22T03:38:59Z) - Regularizing Generative Adversarial Networks under Limited Data [88.57330330305535]
This work proposes a regularization approach for training robust GAN models on limited data.
We show a connection between the regularized loss and an f-divergence called LeCam-divergence, which we find is more robust under limited training data.
arXiv Detail & Related papers (2021-04-07T17:59:06Z) - Adaptive Weighting Scheme for Automatic Time-Series Data Augmentation [79.47771259100674]
We present two sample-adaptive automatic weighting schemes for data augmentation.
We validate our proposed methods on a large, noisy financial dataset and on time-series datasets from the UCR archive.
On the financial dataset, we show that the methods in combination with a trading strategy lead to improvements in annualized returns of over 50$%$, and on the time-series data we outperform state-of-the-art models on over half of the datasets, and achieve similar performance in accuracy on the others.
arXiv Detail & Related papers (2021-02-16T17:50:51Z) - Generalization in Reinforcement Learning by Soft Data Augmentation [11.752595047069505]
SOft Data Augmentation (SODA) is a method that decouples augmentation from policy learning.
We find SODA to significantly advance sample efficiency, generalization, and stability in training over state-of-the-art vision-based RL methods.
arXiv Detail & Related papers (2020-11-26T17:00:34Z) - Generative Data Augmentation for Commonsense Reasoning [75.26876609249197]
G-DAUGC is a novel generative data augmentation method that aims to achieve more accurate and robust learning in the low-resource setting.
G-DAUGC consistently outperforms existing data augmentation methods based on back-translation.
Our analysis demonstrates that G-DAUGC produces a diverse set of fluent training examples, and that its selection and training approaches are important for performance.
arXiv Detail & Related papers (2020-04-24T06:12:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.