Measuring Unintended Memorisation of Unique Private Features in Neural Networks

02/16/2022
by   John Hartley, et al.
1

Neural networks pose a privacy risk to training data due to their propensity to memorise and leak information. Focusing on image classification, we show that neural networks also unintentionally memorise unique features even when they occur only once in training data. An example of a unique feature is a person's name that is accidentally present on a training image. Assuming access to the inputs and outputs of a trained model, the domain of the training data, and knowledge of unique features, we develop a score estimating the model's sensitivity to a unique feature by comparing the KL divergences of the model's output distributions given modified out-of-distribution images. Our results suggest that unique features are memorised by multi-layer perceptrons and convolutional neural networks trained on benchmark datasets, such as MNIST, Fashion-MNIST and CIFAR-10. We find that strategies to prevent overfitting (e.g. early stopping, regularisation, batch normalisation) do not prevent memorisation of unique features. These results imply that neural networks pose a privacy risk to rarely occurring private information. These risks can be more pronounced in healthcare applications if patient information is present in the training data.

READ FULL TEXT

page 1

page 4

page 6

research
05/20/2022

Unintended memorisation of unique features in neural networks

Neural networks pose a privacy risk due to their propensity to memorise ...
research
12/14/2018

Rethinking Layer-wise Feature Amounts in Convolutional Neural Network Architectures

We characterize convolutional neural networks with respect to the relati...
research
05/30/2019

P3SGD: Patient Privacy Preserving SGD for Regularizing Deep CNNs in Pathological Image Classification

Recently, deep convolutional neural networks (CNNs) have achieved great ...
research
04/01/2020

Boosting Deep Hyperspectral Image Classification with Spectral Unmixing

Recent advances in neural networks have made great progress in addressin...
research
06/23/2021

Feature Alignment for Approximated Reversibility in Neural Networks

We introduce feature alignment, a technique for obtaining approximate re...
research
01/28/2019

Complex-Valued Neural Networks for Privacy Protection

This paper proposes a generic method to revise traditional neural networ...
research
06/11/2018

Dual Pattern Learning Networks by Empirical Dual Prediction Risk Minimization

Motivated by the observation that humans can learn patterns from two giv...

Please sign up or login with your details

Forgot password? Click here to reset