Revisiting Priority $k$-Center: Fairness and Outliers
- URL: http://arxiv.org/abs/2103.03337v1
- Date: Thu, 4 Mar 2021 21:15:37 GMT
- Title: Revisiting Priority $k$-Center: Fairness and Outliers
- Authors: Tanvi Bajpai, Deeparnab Chakrabarty, Chandra Chekuri, Maryam Negahbani
- Abstract summary: We develop a framework that yields constant factor approximation algorithms for Priority $k$-Center with outliers.
Our framework extends to generalizations of Priority $k$-Center to matroid and knapsack constraints.
- Score: 5.3898004059026325
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the Priority $k$-Center problem, the input consists of a metric space
$(X,d)$, an integer $k$ and for each point $v \in X$ a priority radius $r(v)$.
The goal is to choose $k$-centers $S \subseteq X$ to minimize $\max_{v \in X}
\frac{1}{r(v)} d(v,S)$. If all $r(v)$'s were uniform, one obtains the classical
$k$-center problem. Plesn\'ik [Plesn\'ik, Disc. Appl. Math. 1987] introduced
this problem and gave a $2$-approximation algorithm matching the best possible
algorithm for vanilla $k$-center. We show how the problem is related to two
different notions of fair clustering [Harris et al., NeurIPS 2018; Jung et al.,
FORC 2020]. Motivated by these developments we revisit the problem and, in our
main technical contribution, develop a framework that yields constant factor
approximation algorithms for Priority $k$-Center with outliers. Our framework
extends to generalizations of Priority $k$-Center to matroid and knapsack
constraints, and as a corollary, also yields algorithms with fairness
guarantees in the lottery model of Harris et al.
Related papers
- Nearly Linear Sparsification of $\ell_p$ Subspace Approximation [47.790126028106734]
A popular approach to cope with the NP-hardness of the $ell_p$ subspace approximation problem is to compute a strong coreset.
We obtain the first algorithm for constructing a strong coreset for $ell_p$ subspace approximation with a nearly optimal dependence on the rank parameter $k$.
Our techniques also lead to the first nearly optimal online strong coresets for $ell_p$ subspace approximation with similar bounds as the offline setting.
arXiv Detail & Related papers (2024-07-03T16:49:28Z) - A Scalable Algorithm for Individually Fair K-means Clustering [77.93955971520549]
We present a scalable algorithm for the individually fair ($p$, $k$)-clustering problem introduced by Jung et al. and Mahabadi et al.
A clustering is then called individually fair if it has centers within distance $delta(x)$ of $x$ for each $xin P$.
We show empirically that not only is our algorithm much faster than prior work, but it also produces lower-cost solutions.
arXiv Detail & Related papers (2024-02-09T19:01:48Z) - Approximating Fair $k$-Min-Sum-Radii in Euclidean Space [1.6369404745833038]
We study the $k$-min-sum-radii problem in Euclidean spaces of arbitrary dimension for the case of constant $k$.
We propose a PTAS for the fair $k$-min-sum-radii problem in Euclidean spaces of arbitrary dimension for the case of constant $k$.
arXiv Detail & Related papers (2023-09-02T06:01:59Z) - Do you know what q-means? [50.045011844765185]
Clustering is one of the most important tools for analysis of large datasets.
We present an improved version of the "$q$-means" algorithm for clustering.
We also present a "dequantized" algorithm for $varepsilon which runs in $Obig(frack2varepsilon2(sqrtkd + log(Nd))big.
arXiv Detail & Related papers (2023-08-18T17:52:12Z) - Improved Learning-augmented Algorithms for k-means and k-medians
Clustering [8.04779839951237]
We consider the problem of clustering in the learning-augmented setting, where we are given a data set in $d$-dimensional Euclidean space.
We propose a deterministic $k$-means algorithm that produces centers with improved bound on clustering cost.
Our algorithm works even when the predictions are not very accurate, i.e. our bound holds for $alpha$ up to $1/2$, an improvement over $alpha$ being at most $1/7$ in the previous work.
arXiv Detail & Related papers (2022-10-31T03:00:11Z) - Corralling a Larger Band of Bandits: A Case Study on Switching Regret
for Linear Bandits [99.86860277006318]
We consider the problem of combining and learning over a set of adversarial algorithms with the goal of adaptively tracking the best one on the fly.
The CORRAL of Agarwal et al. achieves this goal with a regret overhead of order $widetildeO(sqrtd S T)$ where $M$ is the number of base algorithms and $T$ is the time horizon.
Motivated by this issue, we propose a new recipe to corral a larger band of bandit algorithms whose regret overhead has only emphlogarithmic dependence on $M$ as long
arXiv Detail & Related papers (2022-02-12T21:55:44Z) - Improved Approximation Algorithms for Individually Fair Clustering [9.914246432182873]
We present an improved $(16p +varepsilon,3)$-bicriteria approximation for the fair $k$-clustering with $ell_p$-norm cost.
Our approach suggests a reduction from our individually fair clustering to a clustering with a group fairness requirement proposed by Kleindessner et al.
arXiv Detail & Related papers (2021-06-26T15:22:52Z) - Clustering Mixture Models in Almost-Linear Time via List-Decodable Mean
Estimation [58.24280149662003]
We study the problem of list-decodable mean estimation, where an adversary can corrupt a majority of the dataset.
We develop new algorithms for list-decodable mean estimation, achieving nearly-optimal statistical guarantees.
arXiv Detail & Related papers (2021-06-16T03:34:14Z) - Maximizing Determinants under Matroid Constraints [69.25768526213689]
We study the problem of finding a basis $S$ of $M$ such that $det(sum_i in Sv_i v_i v_itop)$ is maximized.
This problem appears in a diverse set of areas such as experimental design, fair allocation of goods, network design, and machine learning.
arXiv Detail & Related papers (2020-04-16T19:16:38Z) - Coreset-based Strategies for Robust Center-type Problems [0.6875312133832077]
We devise coreset-based strategies for the two problems which yield efficient sequential, MapReduce, and Streaming algorithms.
For wide ranges of the parameters $k,zepsilon, D$, we obtain a sequential algorithm with running time linear in $|V|$, and MapReduce/Streaming algorithms with few rounds/passes and substantially sublinear local/working memory.
arXiv Detail & Related papers (2020-02-18T10:04:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.