Maximizing Submodular Functions for Recommendation in the Presence of
Biases
- URL: http://arxiv.org/abs/2305.02806v1
- Date: Wed, 3 May 2023 15:20:00 GMT
- Title: Maximizing Submodular Functions for Recommendation in the Presence of
Biases
- Authors: Anay Mehrotra and Nisheeth K. Vishnoi
- Abstract summary: Subset selection tasks arise in systems and search engines and ask to select a subset of items that maximize the value for the user.
In many applications, inputs have been observed to have social biases that reduce the utility of the output subset.
We show that fairness constraint-based interventions can not only ensure proportional representation but also achieve near-optimal utility in the presence of biases.
- Score: 25.081136190260015
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Subset selection tasks, arise in recommendation systems and search engines
and ask to select a subset of items that maximize the value for the user. The
values of subsets often display diminishing returns, and hence, submodular
functions have been used to model them. If the inputs defining the submodular
function are known, then existing algorithms can be used. In many applications,
however, inputs have been observed to have social biases that reduce the
utility of the output subset. Hence, interventions to improve the utility are
desired. Prior works focus on maximizing linear functions -- a special case of
submodular functions -- and show that fairness constraint-based interventions
can not only ensure proportional representation but also achieve near-optimal
utility in the presence of biases. We study the maximization of a family of
submodular functions that capture functions arising in the aforementioned
applications. Our first result is that, unlike linear functions,
constraint-based interventions cannot guarantee any constant fraction of the
optimal utility for this family of submodular functions. Our second result is
an algorithm for submodular maximization. The algorithm provably outputs
subsets that have near-optimal utility for this family under mild assumptions
and that proportionally represent items from each group. In empirical
evaluation, with both synthetic and real-world data, we observe that this
algorithm improves the utility of the output subset for this family of
submodular functions over baselines.
Related papers
- Learning Submodular Sequencing from Samples [11.528995186765751]
This paper addresses the problem of selecting and ranking items in a sequence to optimize some composite submodular function.
We present an algorithm that achieves an approximation ratio dependent on the curvature of the individual submodular functions.
arXiv Detail & Related papers (2024-09-09T01:33:13Z) - Non-monotone Sequential Submodular Maximization [13.619980548779687]
We aim to select and rank a group of $k$ items from a ground set $V$ such that the weighted assortment of $k$ is maximized.
The results of this research have implications in various fields, including recommendation systems and optimization.
arXiv Detail & Related papers (2023-08-16T19:32:29Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - Neural Estimation of Submodular Functions with Applications to
Differentiable Subset Selection [50.14730810124592]
Submodular functions and variants, through their ability to characterize diversity and coverage, have emerged as a key tool for data selection and summarization.
We propose FLEXSUBNET, a family of flexible neural models for both monotone and non-monotone submodular functions.
arXiv Detail & Related papers (2022-10-20T06:00:45Z) - Using Partial Monotonicity in Submodular Maximization [13.23676270963484]
We show that for many standard submodular algorithms one can prove new approximation guarantees that depend on the monotonicity ratio.
This leads to improved approximation ratios for the common machine learning applications of movie recommendation, quadratic programming and image summarization.
arXiv Detail & Related papers (2022-02-07T10:35:40Z) - Sparsification of Decomposable Submodular Functions [27.070004659301162]
Submodular functions are at the core of many machine learning and data mining tasks.
The number of underlying submodular functions in the original function is so large that we need large amount of time to process it and/or it does not even fit in the main memory.
We introduce the notion of sparsification for decomposable submodular functions whose objective is to obtain an accurate approximation of the original function is a (weighted) sum of only a few submodular functions.
arXiv Detail & Related papers (2022-01-18T20:05:25Z) - Submodular + Concave [53.208470310734825]
It has been well established that first order optimization methods can converge to the maximal objective value of concave functions.
In this work, we initiate the determinant of the smooth functions convex body $$F(x) = G(x) +C(x)$.
This class of functions is an extension of both concave and continuous DR-submodular functions for which no guarantee is known.
arXiv Detail & Related papers (2021-06-09T01:59:55Z) - Planning with Submodular Objective Functions [118.0376288522372]
We study planning with submodular objective functions, where instead of maximizing cumulative reward, the goal is to maximize the value induced by a submodular function.
Our framework subsumes standard planning and submodular objective with cardinality constraints as special cases.
arXiv Detail & Related papers (2020-10-22T16:55:12Z) - Continuous Submodular Function Maximization [91.17492610120324]
Continuous submodularity is a class of functions with a wide spectrum of applications.
We identify several applications of continuous submodular optimization, ranging from influence, MAP for inferences to inferences to field field.
arXiv Detail & Related papers (2020-06-24T04:37:31Z) - From Sets to Multisets: Provable Variational Inference for Probabilistic
Integer Submodular Models [82.95892656532696]
Submodular functions have been studied extensively in machine learning and data mining.
In this work, we propose a continuous DR-submodular extension for integer submodular functions.
We formulate a new probabilistic model which is defined through integer submodular functions.
arXiv Detail & Related papers (2020-06-01T22:20:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.