Federated learning (FedL) is a machine learning (ML) technique utilized to train deep neural networks (DeepNNs) in a distributed way without the need to share data among the federated training clients. FedL was proposed for edge computing and Internet of things (IoT) tasks in which a centralized server was responsible for coordinating and governing the training process. To remove the design limitation implied by the centralized entity, this work proposes two different solutions to decentralize existing FedL algorithms, enabling the application of FedL on networks with arbitrary communication topologies, and thus extending the domain of application of FedL to more complex scenarios and new tasks. Of the two proposed algorithms, one, called FedLCon, is developed based on results from discrete-time weighted average consensus theory and is able to reconstruct the performances of the standard centralized FedL solutions, as also shown by the reported validation tests.

A Weighted Average Consensus Approach for Decentralized Federated Learning / Giuseppi, A.; Manfredi, S.; Pietrabissa, A.. - In: MACHINE INTELLIGENCE RESEARCH. - ISSN 2731-538X. - 19:4(2022), pp. 319-330. [10.1007/s11633-022-1338-z]

A Weighted Average Consensus Approach for Decentralized Federated Learning

Manfredi S.;
2022

Abstract

Federated learning (FedL) is a machine learning (ML) technique utilized to train deep neural networks (DeepNNs) in a distributed way without the need to share data among the federated training clients. FedL was proposed for edge computing and Internet of things (IoT) tasks in which a centralized server was responsible for coordinating and governing the training process. To remove the design limitation implied by the centralized entity, this work proposes two different solutions to decentralize existing FedL algorithms, enabling the application of FedL on networks with arbitrary communication topologies, and thus extending the domain of application of FedL to more complex scenarios and new tasks. Of the two proposed algorithms, one, called FedLCon, is developed based on results from discrete-time weighted average consensus theory and is able to reconstruct the performances of the standard centralized FedL solutions, as also shown by the reported validation tests.
2022
A Weighted Average Consensus Approach for Decentralized Federated Learning / Giuseppi, A.; Manfredi, S.; Pietrabissa, A.. - In: MACHINE INTELLIGENCE RESEARCH. - ISSN 2731-538X. - 19:4(2022), pp. 319-330. [10.1007/s11633-022-1338-z]
File in questo prodotto:
File Dimensione Formato  
A Weighted Average Consensus Approach for Decentralized Federated Learning_2022.pdf

accesso aperto

Licenza: Non specificato
Dimensione 1.11 MB
Formato Adobe PDF
1.11 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11588/902087
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 24
  • ???jsp.display-item.citation.isi??? 21
social impact