FedV: Privacy-Preserving Federated Learning over Vertically Partitioned Data

by   Runhua Xu, et al.

Federated learning (FL) has been proposed to allow collaborative training of machine learning (ML) models among multiple parties where each party can keep its data private. In this paradigm, only model updates, such as model weights or gradients, are shared. Many existing approaches have focused on horizontal FL, where each party has the entire feature set and labels in the training data set. However, many real scenarios follow a vertically-partitioned FL setup, where a complete feature set is formed only when all the datasets from the parties are combined, and the labels are only available to a single party. Privacy-preserving vertical FL is challenging because complete sets of labels and features are not owned by one entity. Existing approaches for vertical FL require multiple peer-to-peer communications among parties, leading to lengthy training times, and are restricted to (approximated) linear models and just two parties. To close this gap, we propose FedV, a framework for secure gradient computation in vertical settings for several widely used ML models such as linear models, logistic regression, and support vector machines. FedV removes the need for peer-to-peer communication among parties by using functional encryption schemes; this allows FedV to achieve faster training times. It also works for larger and changing sets of parties. We empirically demonstrate the applicability for multiple types of ML models and show a reduction of 10 of training time and 80 state-of-the-art approaches.


page 1

page 2

page 3

page 4


Quadratic Functional Encryption for Secure Training in Vertical Federated Learning

Vertical federated learning (VFL) enables the collaborative training of ...

Data Valuation for Vertical Federated Learning: An Information-Theoretic Approach

Federated learning (FL) is a promising machine learning paradigm that en...

Privacy-Preserving Generalized Linear Models using Distributed Block Coordinate Descent

Combining data from varied sources has considerable potential for knowle...

HAFLO: GPU-Based Acceleration for Federated Logistic Regression

In recent years, federated learning (FL) has been widely applied for sup...

Privacy-Preserving Self-Taught Federated Learning for Heterogeneous Data

Many application scenarios call for training a machine learning model am...

Privacy Against Agnostic Inference Attacks in Vertical Federated Learning

A novel form of inference attack in vertical federated learning (VFL) is...

A Communication Efficient Vertical Federated Learning Framework

One critical challenge for applying today's Artificial Intelligence (AI)...

Please sign up or login with your details

Forgot password? Click here to reset