Modelling and Quantifying Membership Information Leakage in Machine Learning

01/29/2020
by   Farhad Farokhi, et al.
4

Machine learning models have been shown to be vulnerable to membership inference attacks, i.e., inferring whether individuals' data have been used for training models. The lack of understanding about factors contributing success of these attacks motivates the need for modelling membership information leakage using information theory and for investigating properties of machine learning models and training algorithms that can reduce membership information leakage. We use conditional mutual information leakage to measure the amount of information leakage from the trained machine learning model about the presence of an individual in the training dataset. We devise an upper bound for this measure of information leakage using Kullback–Leibler divergence that is more amenable to numerical computation. We prove a direct relationship between the Kullback–Leibler membership information leakage and the probability of success for a hypothesis-testing adversary examining whether a particular data record belongs to the training dataset of a machine learning model. We show that the mutual information leakage is a decreasing function of the training dataset size and the regularization weight. We also prove that, if the sensitivity of the machine learning model (defined in terms of the derivatives of the fitness with respect to model parameters) is high, more membership information is potentially leaked. This illustrates that complex models, such as deep neural networks, are more susceptible to membership inference attacks in comparison to simpler models with fewer degrees of freedom. We show that the amount of the membership information leakage is reduced by O(log^1/2(δ^-1)ϵ^-1) when using Gaussian (ϵ,δ)-differentially-private additive noises.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/18/2016

Membership Inference Attacks against Machine Learning Models

We quantitatively investigate how machine learning models leak informati...
research
09/17/2020

An Extension of Fano's Inequality for Characterizing Model Susceptibility to Membership Inference Attacks

Deep neural networks have been shown to be vulnerable to membership infe...
research
05/09/2021

Bounding Information Leakage in Machine Learning

Machine Learning services are being deployed in a large range of applica...
research
02/17/2020

Data and Model Dependencies of Membership Inference Attack

Machine Learning (ML) techniques are used by most data-driven organisati...
research
10/12/2020

Quantifying Membership Privacy via Information Leakage

Machine learning models are known to memorize the unique properties of i...
research
11/18/2019

Privacy Leakage Avoidance with Switching Ensembles

We consider membership inference attacks, one of the main privacy issues...
research
10/05/2019

Characterizing Membership Privacy in Stochastic Gradient Langevin Dynamics

Bayesian deep learning is recently regarded as an intrinsic way to chara...

Please sign up or login with your details

Forgot password? Click here to reset