Federated Learning convergence on IID features via optimized local model parameters

Shan Ullah, Deok Hwan Kim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

5 Scopus citations

Abstract

Federated learning enables collaborative training of global models without sharing local training data. In this paper, we propose a method to enhance the convergence of federated learning algorithm by modifying the training method at the client's side, such as we have deployed two different methods at local models, i) filter out batch normalization parameters and remove them from weight decay, and ii) zero initialize the last batch normalization in each residual branch (in case of resnet) so that the residual branch starts with zeros and each residual block behaves as identity. The following two methods were previously proposed for distributed training. In this paper, we propose to deploy the method in federated learning to enhance the convergence of the federated aggregation method. The proposed method is implemented at local model training and has shown significant improvements in the convergence, keeping the same rounds of federated learning. We have rigorously evaluated the efficacy of the proposed algorithm using a range of image classification models such as Resnet (-18, -50) and MobileNetv2 under the CIFAR-10 dataset. For all experiments, we have used an independent and identically distributed (IID) configuration for distributing the data among contributing clients and fedavg as a federated aggregation algorithm. We have deployed our code based on the open-source library 'Flower' for federated learning and communication. We have presented detailed results on various performance metrices such convergence under pretrained networks and without pretrained networks with different number of federated rounds, hyperparameter optimization, impact on different models to provide generability. The proposed algorithm outperforms state-of-the art algorithms in terms of convergence for federated learning by increasing up to 1.5% accuracy in limited federated rounds.

Original languageEnglish
Title of host publicationProceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022
EditorsHerwig Unger, Young-Kuk Kim, Eenjun Hwang, Sung-Bae Cho, Stephan Pareigis, Kyamakya Kyandoghere, Young-Guk Ha, Jinho Kim, Atsuyuki Morishima, Christian Wagner, Hyuk-Yoon Kwon, Yang-Sae Moon, Carson Leung
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages92-95
Number of pages4
ISBN (Electronic)9781665421973
DOIs
StatePublished - 2022
Event2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022 - Daegu, Korea, Republic of
Duration: 17 Jan 202220 Jan 2022

Publication series

NameProceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022

Conference

Conference2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022
Country/TerritoryKorea, Republic of
CityDaegu
Period17/01/2220/01/22

Bibliographical note

Publisher Copyright:
© 2022 IEEE.

Keywords

  • IID
  • batch normalization
  • deep learning
  • federated learning

Fingerprint

Dive into the research topics of 'Federated Learning convergence on IID features via optimized local model parameters'. Together they form a unique fingerprint.

Cite this