DFedADMM: Dual Constraints Controlled Model Inconsistency for Decentralized Federated Learning

08/16/2023
by   Qinglun Li, et al.
0

To address the communication burden issues associated with federated learning (FL), decentralized federated learning (DFL) discards the central server and establishes a decentralized communication network, where each client communicates only with neighboring clients. However, existing DFL methods still suffer from two major challenges: local inconsistency and local heterogeneous overfitting, which have not been fundamentally addressed by existing DFL methods. To tackle these issues, we propose novel DFL algorithms, DFedADMM and its enhanced version DFedADMM-SAM, to enhance the performance of DFL. The DFedADMM algorithm employs primal-dual optimization (ADMM) by utilizing dual variables to control the model inconsistency raised from the decentralized heterogeneous data distributions. The DFedADMM-SAM algorithm further improves on DFedADMM by employing a Sharpness-Aware Minimization (SAM) optimizer, which uses gradient perturbations to generate locally flat models and searches for models with uniformly low loss values to mitigate local heterogeneous overfitting. Theoretically, we derive convergence rates of 𝒪(1/√(KT)+1/KT(1-ψ)^2) and 𝒪(1/√(KT)+1/KT(1-ψ)^2+ 1/T^3/2K^1/2) in the non-convex setting for DFedADMM and DFedADMM-SAM, respectively, where 1 - ψ represents the spectral gap of the gossip matrix. Empirically, extensive experiments on MNIST, CIFAR10 and CIFAR100 datesets demonstrate that our algorithms exhibit superior performance in terms of both generalization and convergence speed compared to existing state-of-the-art (SOTA) optimizers in DFL.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2023

Improving the Model Consistency of Decentralized Federated Learning

To mitigate the privacy leakages and communication burdens of Federated ...
research
08/31/2023

Robust Networked Federated Learning for Localization

This paper addresses the problem of localization, which is inherently no...
research
12/03/2022

Beyond ADMM: A Unified Client-variance-reduced Adaptive Federated Learning Framework

As a novel distributed learning paradigm, federated learning (FL) faces ...
research
10/25/2022

SWIFT: Rapid Decentralized Federated Learning via Wait-Free Model Communication

The decentralized Federated Learning (FL) setting avoids the role of a p...
research
06/22/2022

: Calibrating Global and Local Models via Federated Learning Beyond Consensus

In federated learning (FL), the objective of collaboratively learning a ...
research
05/08/2022

Communication Compression for Decentralized Learning with Operator Splitting Methods

In decentralized learning, operator splitting methods using a primal-dua...
research
09/19/2022

Heterogeneous Federated Learning on a Graph

Federated learning, where algorithms are trained across multiple decentr...

Please sign up or login with your details

Forgot password? Click here to reset