Scalable and Efficient Training of Large Convolutional Neural Networks with Differential Privacy

05/21/2022
by   Zhiqi Bu, et al.
0

Large convolutional neural networks (CNN) can be difficult to train in the differentially private (DP) regime, since the optimization algorithms require a computationally expensive operation, known as the per-sample gradient clipping. We propose an efficient and scalable implementation of this clipping on convolutional layers, termed as the mixed ghost clipping, that significantly eases the private training in terms of both time and space complexities, without affecting the accuracy. The improvement in efficiency is rigorously studied through the first complexity analysis for the mixed ghost clipping and existing DP training algorithms. Extensive experiments on vision classification tasks, with large ResNet, VGG, and Vision Transformers, demonstrate that DP training with mixed ghost clipping adds 1∼ 10% memory overhead and <2× slowdown to the standard non-private training. Specifically, when training VGG19 on CIFAR10, the mixed ghost clipping is 3× faster than state-of-the-art Opacus library with 18× larger maximum batch size. To emphasize the significance of efficient DP training on convolutional layers, we achieve 96.7% accuracy on CIFAR10 and 83.0% on CIFAR100 at ϵ=1 using BEiT, while the previous best results are 94.8% and 67.4%, respectively. We open-source a privacy engine (<https://github.com/JialinMao/private_CNN>) that implements DP training of CNN with a few lines of code.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/30/2022

Differentially Private Optimization on Large Model at Small Cost

Differentially private (DP) optimization is the standard paradigm to lea...
research
09/30/2022

Differentially Private Bias-Term only Fine-tuning of Foundation Models

We study the problem of differentially private (DP) fine-tuning of large...
research
01/30/2023

Equivariant Differentially Private Deep Learning

The formal privacy guarantee provided by Differential Privacy (DP) bound...
research
06/19/2020

Robust Differentially Private Training of Deep Neural Networks

Differentially private stochastic gradient descent (DPSGD) is a variatio...
research
09/12/2023

Exploring the Benefits of Differentially Private Pre-training and Parameter-Efficient Fine-tuning for Table Transformers

For machine learning with tabular data, Table Transformer (TabTransforme...
research
06/09/2023

Differentially Private Sharpness-Aware Training

Training deep learning models with differential privacy (DP) results in ...
research
05/16/2021

Dynamic Pooling Improves Nanopore Base Calling Accuracy

In nanopore sequencing, electrical signal is measured as DNA molecules p...

Please sign up or login with your details

Forgot password? Click here to reset