Mitigating Unintended Memorization in Language Models via Alternating Teaching

10/13/2022
by   Zhe Liu, et al.
0

Recent research has shown that language models have a tendency to memorize rare or unique sequences in the training corpora which can thus leak sensitive attributes of user data. We employ a teacher-student framework and propose a novel approach called alternating teaching to mitigate unintended memorization in sequential modeling. In our method, multiple teachers are trained on disjoint training sets whose privacy one wishes to protect, and teachers' predictions supervise the training of a student model in an alternating manner at each time step. Experiments on LibriSpeech datasets show that the proposed method achieves superior privacy-preserving results than other counterparts. In comparison with no prevention for unintended memorization, the overall utility loss is small when training records are sufficient.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2019

Not Just Cloud Privacy: Protecting Client Privacy in Teacher-Student Learning

Ensuring the privacy of sensitive data used to train modern machine lear...
research
04/05/2020

Private Knowledge Transfer via Model Distillation with Generative Adversarial Networks

The deployment of deep learning applications has to address the growing ...
research
07/09/2020

Learning to Teach with Deep Interactions

Machine teaching uses a meta/teacher model to guide the training of a st...
research
06/14/2023

Protecting User Privacy in Remote Conversational Systems: A Privacy-Preserving framework based on text sanitization

Large Language Models (LLMs) are gaining increasing attention due to the...
research
12/16/2022

Planting and Mitigating Memorized Content in Predictive-Text Language Models

Language models are widely deployed to provide automatic text completion...
research
09/12/2023

Recovering from Privacy-Preserving Masking with Large Language Models

Model adaptation is crucial to handle the discrepancy between proxy trai...
research
05/02/2023

Mitigating Approximate Memorization in Language Models via Dissimilarity Learned Policy

Large Language models (LLMs) are trained on large amounts of data, which...

Please sign up or login with your details

Forgot password? Click here to reset