Doubly Constrained Fair Clustering

by   John Dickerson, et al.

The remarkable attention which fair clustering has received in the last few years has resulted in a significant number of different notions of fairness. Despite the fact that these notions are well-justified, they are often motivated and studied in a disjoint manner where one fairness desideratum is considered exclusively in isolation from the others. This leaves the understanding of the relations between different fairness notions as an important open problem in fair clustering. In this paper, we take the first step in this direction. Specifically, we consider the two most prominent demographic representation fairness notions in clustering: (1) Group Fairness (GF), where the different demographic groups are supposed to have close to population-level representation in each cluster and (2) Diversity in Center Selection (DS), where the selected centers are supposed to have close to population-level representation of each group. We show that given a constant approximation algorithm for one constraint (GF or DS only) we can obtain a constant approximation solution that satisfies both constraints simultaneously. Interestingly, we prove that any given solution that satisfies the GF constraint can always be post-processed at a bounded degradation to the clustering cost to additionally satisfy the DS constraint while the reverse is not true. Furthermore, we show that both GF and DS are incompatible (having an empty feasibility set in the worst case) with a collection of other distance-based fairness notions. Finally, we carry experiments to validate our theoretical findings.


page 1

page 2

page 3

page 4


Fair Labeled Clustering

Numerous algorithms have been produced for the fundamental problem of cl...

Fair Representation: Guaranteeing Approximate Multiple Group Fairness for Unknown Tasks

Motivated by scenarios where data is used for diverse prediction tasks, ...

Fair k-Center Clustering for Data Summarization

In data summarization we want to choose k prototypes in order to summari...

Fair clustering via equitable group representations

What does it mean for a clustering to be fair? One popular approach seek...

Fair Prediction with Endogenous Behavior

There is increasing regulatory interest in whether machine learning algo...

An example of prediction which complies with Demographic Parity and equalizes group-wise risks in the context of regression

Let (X, S, Y) ∈ℝ^p ×{1, 2}×ℝ be a triplet following some joint distribut...

A Fair and Memory/Time-efficient Hashmap

There is a large amount of work constructing hashmaps to minimize the nu...

Please sign up or login with your details

Forgot password? Click here to reset