A Pairwise Fair and Community-preserving Approach to k-Center Clustering
- URL: http://arxiv.org/abs/2007.07384v1
- Date: Tue, 14 Jul 2020 22:32:27 GMT
- Title: A Pairwise Fair and Community-preserving Approach to k-Center Clustering
- Authors: Brian Brubach, Darshan Chakrabarti, John P. Dickerson, Samir Khuller,
Aravind Srinivasan, Leonidas Tsepenekas
- Abstract summary: Clustering is a foundational problem in machine learning with numerous applications.
We define two new types of fairness in the clustering setting, pairwise fairness and community preservation.
- Score: 34.386585230600716
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Clustering is a foundational problem in machine learning with numerous
applications. As machine learning increases in ubiquity as a backend for
automated systems, concerns about fairness arise. Much of the current
literature on fairness deals with discrimination against protected classes in
supervised learning (group fairness). We define a different notion of fair
clustering wherein the probability that two points (or a community of points)
become separated is bounded by an increasing function of their pairwise
distance (or community diameter). We capture the situation where data points
represent people who gain some benefit from being clustered together.
Unfairness arises when certain points are deterministically separated, either
arbitrarily or by someone who intends to harm them as in the case of
gerrymandering election districts. In response, we formally define two new
types of fairness in the clustering setting, pairwise fairness and community
preservation. To explore the practicality of our fairness goals, we devise an
approach for extending existing $k$-center algorithms to satisfy these fairness
constraints. Analysis of this approach proves that reasonable approximations
can be achieved while maintaining fairness. In experiments, we compare the
effectiveness of our approach to classical $k$-center algorithms/heuristics and
explore the tradeoff between optimal clustering and fairness.
Related papers
- Proportionally Representative Clustering [17.5359577544947]
We propose a new axiom proportionally representative fairness'' (PRF) that is designed for clustering problems.
Our fairness concept is not satisfied by existing fair clustering algorithms.
Our algorithm for the unconstrained setting is also the first known-time approximation algorithm for the well-studied Proportional Fairness (PF) axiom.
arXiv Detail & Related papers (2023-04-27T02:01:24Z) - DualFair: Fair Representation Learning at Both Group and Individual
Levels via Contrastive Self-supervision [73.80009454050858]
This work presents a self-supervised model, called DualFair, that can debias sensitive attributes like gender and race from learned representations.
Our model jointly optimize for two fairness criteria - group fairness and counterfactual fairness.
arXiv Detail & Related papers (2023-03-15T07:13:54Z) - Fair Correlation Clustering in Forests [8.810926150873991]
A clustering is said to be fair, if each cluster has the same distribution of manifestations of a sensitive attribute as the whole input set.
This is motivated by various applications where the objects to be clustered have sensitive attributes that should not be over- or underrepresented.
We consider restricted graph classes which allow us to characterize the distributions of sensitive attributes for which this form of fairness is tractable.
arXiv Detail & Related papers (2023-02-22T11:27:06Z) - Fairness in Matching under Uncertainty [78.39459690570531]
algorithmic two-sided marketplaces have drawn attention to the issue of fairness in such settings.
We axiomatize a notion of individual fairness in the two-sided marketplace setting which respects the uncertainty in the merits.
We design a linear programming framework to find fair utility-maximizing distributions over allocations.
arXiv Detail & Related papers (2023-02-08T00:30:32Z) - Practical Approaches for Fair Learning with Multitype and Multivariate
Sensitive Attributes [70.6326967720747]
It is important to guarantee that machine learning algorithms deployed in the real world do not result in unfairness or unintended social consequences.
We introduce FairCOCCO, a fairness measure built on cross-covariance operators on reproducing kernel Hilbert Spaces.
We empirically demonstrate consistent improvements against state-of-the-art techniques in balancing predictive power and fairness on real-world datasets.
arXiv Detail & Related papers (2022-11-11T11:28:46Z) - How Robust is Your Fairness? Evaluating and Sustaining Fairness under
Unseen Distribution Shifts [107.72786199113183]
We propose a novel fairness learning method termed CUrvature MAtching (CUMA)
CUMA achieves robust fairness generalizable to unseen domains with unknown distributional shifts.
We evaluate our method on three popular fairness datasets.
arXiv Detail & Related papers (2022-07-04T02:37:50Z) - Fair Clustering Under a Bounded Cost [33.50262066253557]
Clustering is a fundamental unsupervised learning problem where a dataset is partitioned into clusters that consist of nearby points in a metric space.
A recent variant, fair clustering, associates a color with each point representing its group membership and requires that each color has (approximately) equal representation in each cluster to satisfy group fairness.
We consider two fairness objectives: the group utilitarian objective and the group egalitarian objective, as well as the group leximin objective which generalizes the group egalitarian objective.
arXiv Detail & Related papers (2021-06-14T08:47:36Z) - MultiFair: Multi-Group Fairness in Machine Learning [52.24956510371455]
We study multi-group fairness in machine learning (MultiFair)
We propose a generic end-to-end algorithmic framework to solve it.
Our proposed framework is generalizable to many different settings.
arXiv Detail & Related papers (2021-05-24T02:30:22Z) - Distributional Individual Fairness in Clustering [7.303841123034983]
We introduce a framework for assigning individuals, embedded in a metric space, to probability distributions over a bounded number of cluster centers.
We provide an algorithm for clustering with $p$-norm objective and individual fairness constraints with provable approximation guarantee.
arXiv Detail & Related papers (2020-06-22T20:02:09Z) - Fair Hierarchical Clustering [92.03780518164108]
We define a notion of fairness that mitigates over-representation in traditional clustering.
We show that our algorithms can find a fair hierarchical clustering, with only a negligible loss in the objective.
arXiv Detail & Related papers (2020-06-18T01:05:11Z) - A Notion of Individual Fairness for Clustering [22.288902523866867]
A common distinction in fair machine learning, in particular in fair classification, is between group fairness and individual fairness.
In this paper, we propose a natural notion of individual fairness for clustering. Our notion asks that every data point, on average, is closer to the points in its own cluster than to the points in any other cluster.
arXiv Detail & Related papers (2020-06-08T21:41:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.