Federated learning (FL) can be used to distribute machine learning (ML) tasks across edge and Internet of Things (IoT) devices with limited resources. FL provides an alternative and much more practical solution to classical artificial intelligence (AI), which requires moving large data volumes to energy-hungry data centers. On the other hand, sustainability of FL processes should be accurately quantified as limiting energy consumption might require sacrificing accuracy. This article proposes a framework for real-time monitoring of energy and green house gas (GHG) emissions (carbon footprints) of FL systems. The framework is developed for both classical FL policies relying on the parameter server and emerging fully decentralized ones. The proposed approach considers, for the first time, the impact of ML model quantization and sparsification on the energy/carbon budget while also discussing novel gradient tracking (GT) FL strategies that are robust to data heterogeneity but require higher communication bandwidth. General guidelines for energy-efficient designs are discussed based on several case studies on real datasets. This article quantifies the energy footprint of continual FL processes that implement periodic adaptation on new data as foreseen by emerging IoT industry verticals. Results show that centralized FL is advantageous when strict carbon budgets are imposed or energy-inefficient (< 50 Kbit/Joule) communication protocols are adopted. GT mechanisms are to be preferred in heterogeneous data environments and decentralized setups. Finally, using FL for continual model fine-tuning provides large energy savings (> 80%), provided the ML model compression is properly tuned.
A Close Look at the Communication Efficiency and the Energy Footprints of Robust Federated Learning in Industrial IoT
Barbieri, Luca;Kianoush, Sanaz;Nicoli, Monica;Savazzi, Stefano
2025-01-01
Abstract
Federated learning (FL) can be used to distribute machine learning (ML) tasks across edge and Internet of Things (IoT) devices with limited resources. FL provides an alternative and much more practical solution to classical artificial intelligence (AI), which requires moving large data volumes to energy-hungry data centers. On the other hand, sustainability of FL processes should be accurately quantified as limiting energy consumption might require sacrificing accuracy. This article proposes a framework for real-time monitoring of energy and green house gas (GHG) emissions (carbon footprints) of FL systems. The framework is developed for both classical FL policies relying on the parameter server and emerging fully decentralized ones. The proposed approach considers, for the first time, the impact of ML model quantization and sparsification on the energy/carbon budget while also discussing novel gradient tracking (GT) FL strategies that are robust to data heterogeneity but require higher communication bandwidth. General guidelines for energy-efficient designs are discussed based on several case studies on real datasets. This article quantifies the energy footprint of continual FL processes that implement periodic adaptation on new data as foreseen by emerging IoT industry verticals. Results show that centralized FL is advantageous when strict carbon budgets are imposed or energy-inefficient (< 50 Kbit/Joule) communication protocols are adopted. GT mechanisms are to be preferred in heterogeneous data environments and decentralized setups. Finally, using FL for continual model fine-tuning provides large energy savings (> 80%), provided the ML model compression is properly tuned.| File | Dimensione | Formato | |
|---|---|---|---|
|
A_Close_Look_at_the_Communication_Efficiency_and_the_Energy_Footprints_of_Robust_Federated_Learning_in_Industrial_IoT.pdf
accesso aperto
Descrizione: Full paper
:
Publisher’s version
Dimensione
3.67 MB
Formato
Adobe PDF
|
3.67 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


