Characterizing and Optimizing End-to-End Systems for Private Inference

07/14/2022
by   Karthik Garimella, et al.
0

Increasing privacy concerns have given rise to Private Inference (PI). In PI, both the client's personal data and the service provider's trained model are kept confidential. State-of-the-art PI protocols combine several cryptographic primitives: Homomorphic Encryption (HE), Secret Sharing (SS), Garbled Circuits (GC), and Oblivious Transfer (OT). Today, PI remains largely arcane and too slow for practical use, despite the need and recent performance improvements. This paper addresses PI's shortcomings with a detailed characterization of a standard high-performance protocol to build foundational knowledge and intuition in the systems community. The characterization pinpoints all sources of inefficiency – compute, communication, and storage. A notable aspect of this work is the use of inference request arrival rates rather than studying individual inferences in isolation. Prior to this work, and without considering arrival rate, it has been assumed that PI pre-computations can be handled offline and their overheads ignored. We show this is not the case. The offline costs in PI are so high that they are often incurred online, as there is insufficient downtime to hide pre-compute latency. We further propose three optimizations to address the computation (layer-parallel HE), communication (wireless slot allocation), and storage (Client-Garbler) overheads leveraging insights from our characterization. Compared to the state-of-the-art PI protocol, the optimizations provide a total PI speedup of 1.8×, with the ability to sustain inference requests up to a 2.24× greater rate.

READ FULL TEXT

page 4

page 7

page 10

research
11/04/2021

CryptoNite: Revealing the Pitfalls of End-to-End Private Inference at Scale

The privacy concerns of providing deep learning inference as a service h...
research
05/13/2022

Impala: Low-Latency, Communication-Efficient Private Deep Learning Inference

This paper proposes Impala, a new cryptographic protocol for private inf...
research
06/15/2021

Circa: Stochastic ReLUs for Private Deep Learning

The simultaneous rise of machine learning as a service and concerns over...
research
07/09/2023

Towards Fast and Scalable Private Inference

Privacy and security have rapidly emerged as first order design constrai...
research
07/25/2020

SOTERIA: In Search of Efficient Neural Networks for Private Inference

ML-as-a-service is gaining popularity where a cloud server hosts a train...
research
03/02/2021

DeepReDuce: ReLU Reduction for Fast Private Inference

The recent rise of privacy concerns has led researchers to devise method...
research
02/27/2022

Split HE: Fast Secure Inference Combining Split Learning and Homomorphic Encryption

This work presents a novel protocol for fast secure inference of neural ...

Please sign up or login with your details

Forgot password? Click here to reset