Breaking Fair Binary Classification with Optimal Flipping Attacks

04/12/2022
by   Changhun Jo, et al.
0

Minimizing risk with fairness constraints is one of the popular approaches to learning a fair classifier. Recent works showed that this approach yields an unfair classifier if the training set is corrupted. In this work, we study the minimum amount of data corruption required for a successful flipping attack. First, we find lower/upper bounds on this quantity and show that these bounds are tight when the target model is the unique unconstrained risk minimizer. Second, we propose a computationally efficient data poisoning attack algorithm that can compromise the performance of fair learning algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/13/2022

FARE: Provably Fair Representation Learning

Fair representation learning (FRL) is a popular class of methods aiming ...
research
02/11/2021

Fairness-Aware Learning from Corrupted Data

Addressing fairness concerns about machine learning models is a crucial ...
research
03/06/2018

A Reductions Approach to Fair Classification

We present a systematic approach for achieving fairness in a binary clas...
research
01/10/2018

Optimal functional supervised classification with separation condition

We consider the binary supervised classification problem with the Gaussi...
research
05/23/2019

Binary Classification with Bounded Abstention Rate

We consider the problem of binary classification with abstention in the ...
research
10/18/2022

Towards Fair Classification against Poisoning Attacks

Fair classification aims to stress the classification models to achieve ...
research
06/30/2020

Model-Targeted Poisoning Attacks: Provable Convergence and Certified Bounds

Machine learning systems that rely on training data collected from untru...

Please sign up or login with your details

Forgot password? Click here to reset