Federated Submodel Averaging

09/16/2021
by   Yucheng Ding, et al.
0

We study practical data characteristics underlying federated learning, where non-i.i.d. data from clients have sparse features, and a certain client's local data normally involves only a small part of the full model, called a submodel. Due to data sparsity, the classical federated averaging (FedAvg) algorithm or its variants will be severely slowed down, because when updating the global model, each client's zero update of the full model excluding its submodel is inaccurately aggregated. Therefore, we propose federated submodel averaging (FedSubAvg), ensuring that the expectation of the global update of each model parameter is equal to the average of the local updates of the clients who involve it. We theoretically proved the convergence rate of FedSubAvg by deriving an upper bound under a new metric called the element-wise gradient norm. In particular, this new metric can characterize the convergence of federated optimization over sparse data, while the conventional metric of squared gradient norm used in FedAvg and its variants cannot. We extensively evaluated FedSubAvg over both public and industrial datasets. The evaluation results demonstrate that FedSubAvg significantly outperforms FedAvg and its variants.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2022

Semi-Decentralized Federated Learning with Collaborative Relaying

We present a semi-decentralized federated learning algorithm wherein cli...
research
01/28/2022

Gradient Masked Averaging for Federated Learning

Federated learning is an emerging paradigm that permits a large number o...
research
02/18/2020

Distributed Optimization over Block-Cyclic Data

We consider practical data characteristics underlying federated learning...
research
12/17/2021

From Deterioration to Acceleration: A Calibration Approach to Rehabilitating Step Asynchronism in Federated Optimization

In the setting of federated optimization, where a global model is aggreg...
research
11/28/2021

Federated Gaussian Process: Convergence, Automatic Personalization and Multi-fidelity Modeling

In this paper, we propose : a Federated Gaussian process (𝒢𝒫) regression...
research
12/09/2021

On Convergence of Federated Averaging Langevin Dynamics

We propose a federated averaging Langevin algorithm (FA-LD) for uncertai...
research
06/13/2022

Accelerating Federated Learning via Sampling Anchor Clients with Large Batches

Using large batches in recent federated learning studies has improved co...

Please sign up or login with your details

Forgot password? Click here to reset