Data Optimisation for a Deep Learning Recommender System
- URL: http://arxiv.org/abs/2106.11218v1
- Date: Mon, 21 Jun 2021 16:05:37 GMT
- Title: Data Optimisation for a Deep Learning Recommender System
- Authors: Gustav Hertz, Sandhya Sachidanandan, Bal\'azs T\'oth, Emil S.
J{\o}rgensen and Martin Tegn\'er
- Abstract summary: This paper advocates privacy preserving requirements on collection of user data for recommender systems.
First, we ask if restrictions on data collection will hurt test quality of RNN-based recommendations.
Second, we ask if we can improve the quality under minimal data by using secondary data sources.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper advocates privacy preserving requirements on collection of user
data for recommender systems. The purpose of our study is twofold. First, we
ask if restrictions on data collection will hurt test quality of RNN-based
recommendations. We study how validation performance depends on the available
amount of training data. We use a combination of top-K accuracy, catalog
coverage and novelty for this purpose, since good recommendations for the user
is not necessarily captured by a traditional accuracy metric. Second, we ask if
we can improve the quality under minimal data by using secondary data sources.
We propose knowledge transfer for this purpose and construct a representation
to measure similarities between purchase behaviour in data. This to make
qualified judgements of which source domain will contribute the most. Our
results show that (i) there is a saturation in test performance when training
size is increased above a critical point. We also discuss the interplay between
different performance metrics, and properties of data. Moreover, we demonstrate
that (ii) our representation is meaningful for measuring purchase behaviour. In
particular, results show that we can leverage secondary data to improve
validation performance if we select a relevant source domain according to our
similarly measure.
Related papers
- Beyond Models! Explainable Data Valuation and Metric Adaption for Recommendation [10.964035199849125]
Current methods employ data valuation to discern high-quality data from low-quality data.
We propose an explainable and versatile framework DVR which can enhance the efficiency of data utilization tailored to any requirements.
Our framework achieves up to 34.7% improvements over existing methods in terms of representative NDCG metric.
arXiv Detail & Related papers (2025-02-12T12:01:08Z) - Reward-Augmented Data Enhances Direct Preference Alignment of LLMs [63.32585910975191]
We introduce reward-conditioned Large Language Models (LLMs) that learn from the entire spectrum of response quality within the dataset.
We propose an effective yet simple data relabeling method that conditions the preference pairs on quality scores to construct a reward-augmented dataset.
arXiv Detail & Related papers (2024-10-10T16:01:51Z) - Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback [110.16220825629749]
Learning from preference feedback has emerged as an essential step for improving the generation quality and performance of modern language models.
In this work, we identify four core aspects of preference-based learning: preference data, learning algorithm, reward model, and policy training prompts.
Our findings indicate that all aspects are important for performance, with better preference data leading to the largest improvements.
arXiv Detail & Related papers (2024-06-13T16:17:21Z) - Exploring the Mystery of Influential Data for Mathematical Reasoning [127.61978092016228]
We propose a Quality-aware Diverse Selection (QaDS) strategy for mathematical reasoning.
A comparison with other selection strategies validates the superiority of QaDS.
With OpenMathMix, we achieve a state-of-the-art 48.8% accuracy on MATH with 7B base model.
arXiv Detail & Related papers (2024-04-01T12:01:06Z) - Post-Training Attribute Unlearning in Recommender Systems [37.67195112898097]
Existing studies predominantly use training data, i.e., model inputs, as unlearning target.
We name this unseen information as textitattribute and treat it as unlearning target.
To protect the sensitive attribute of users, Attribute Unlearning (AU) aims to make target attributes indistinguishable.
arXiv Detail & Related papers (2024-03-11T14:02:24Z) - LESS: Selecting Influential Data for Targeted Instruction Tuning [64.78894228923619]
We propose LESS, an efficient algorithm to estimate data influences and perform Low-rank gradiEnt Similarity Search for instruction data selection.
We show that training on a LESS-selected 5% of the data can often outperform training on the full dataset across diverse downstream tasks.
Our method goes beyond surface form cues to identify data that the necessary reasoning skills for the intended downstream application.
arXiv Detail & Related papers (2024-02-06T19:18:04Z) - LAVA: Data Valuation without Pre-Specified Learning Algorithms [20.578106028270607]
We introduce a new framework that can value training data in a way that is oblivious to the downstream learning algorithm.
We develop a proxy for the validation performance associated with a training set based on a non-conventional class-wise Wasserstein distance between training and validation sets.
We show that the distance characterizes the upper bound of the validation performance for any given model under certain Lipschitz conditions.
arXiv Detail & Related papers (2023-04-28T19:05:16Z) - Understanding or Manipulation: Rethinking Online Performance Gains of
Modern Recommender Systems [38.75457258877731]
We present a framework for benchmarking the degree of manipulations of recommendation algorithms.
We find that a high online click-through rate does not necessarily mean a better understanding of user initial preference.
We advocate that future recommendation algorithm studies should be treated as an optimization problem with constrained user preference manipulations.
arXiv Detail & Related papers (2022-10-11T17:56:55Z) - Recommendation Systems with Distribution-Free Reliability Guarantees [83.80644194980042]
We show how to return a set of items rigorously guaranteed to contain mostly good items.
Our procedure endows any ranking model with rigorous finite-sample control of the false discovery rate.
We evaluate our methods on the Yahoo! Learning to Rank and MSMarco datasets.
arXiv Detail & Related papers (2022-07-04T17:49:25Z) - Learning PAC-Bayes Priors for Probabilistic Neural Networks [32.01506699213665]
Recent works have investigated deep learning models trained by optimising PAC-Bayes bounds, with priors that are learnt on subsets of the data.
We ask what is the optimal amount of data which should be allocated for building the prior and show that the optimum may be dataset dependent.
arXiv Detail & Related papers (2021-09-21T16:27:42Z) - Improving Multi-Turn Response Selection Models with Complementary
Last-Utterance Selection by Instance Weighting [84.9716460244444]
We consider utilizing the underlying correlation in the data resource itself to derive different kinds of supervision signals.
We conduct extensive experiments in two public datasets and obtain significant improvement in both datasets.
arXiv Detail & Related papers (2020-02-18T06:29:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.