Personalized Federated Learning (PFL) tools have been recently applied in Anomaly Detection (AD) setups to accurately monitor complex industrial systems under data heterogeneity while complying with strict privacy regulations. PFL techniques integrating transformer models in AD setups are still overlooked even though they provide outstanding performances that are hardly matched by other Neural Network (NN) architectures. This paper thus focuses on developing transformer-based PFL techniques in AD contexts to improve AD accuracy under data heterogeneity. Specifically, we propose decoupling the FL optimization process in a layer-wise manner by carefully selecting which model fragments are learned collaboratively and which are personalized (i.e., trained without cooperation). We refer to our proposed methodology as Layer-Wise Personalized FL (LPFL). The developed approach is evaluated with four design choices for selecting the model layers tailored according to the peculiar architecture of transformer NNs (e.g., the self-attention mechanism). Experimental results on four widely-adopted AD datasets highlight that the self-attention mechanism should always be learned collaboratively while all other trainable parameters should be personalized. Adopting such a choice boosts AD accuracy and reduces the communication overhead by up to 16% and 52%, respectively, compared to other personalization choices, standard FL policies, and individual training strategies.

A Layer-Wise Personalization Approach for Transformer-Based Federated Anomaly Detection

Brambilla, Mattia;Roveri, Manuel
2024-01-01

Abstract

Personalized Federated Learning (PFL) tools have been recently applied in Anomaly Detection (AD) setups to accurately monitor complex industrial systems under data heterogeneity while complying with strict privacy regulations. PFL techniques integrating transformer models in AD setups are still overlooked even though they provide outstanding performances that are hardly matched by other Neural Network (NN) architectures. This paper thus focuses on developing transformer-based PFL techniques in AD contexts to improve AD accuracy under data heterogeneity. Specifically, we propose decoupling the FL optimization process in a layer-wise manner by carefully selecting which model fragments are learned collaboratively and which are personalized (i.e., trained without cooperation). We refer to our proposed methodology as Layer-Wise Personalized FL (LPFL). The developed approach is evaluated with four design choices for selecting the model layers tailored according to the peculiar architecture of transformer NNs (e.g., the self-attention mechanism). Experimental results on four widely-adopted AD datasets highlight that the self-attention mechanism should always be learned collaboratively while all other trainable parameters should be personalized. Adopting such a choice boosts AD accuracy and reduces the communication overhead by up to 16% and 52%, respectively, compared to other personalization choices, standard FL policies, and individual training strategies.
2024
2024 2nd International Conference on Federated Learning Technologies and Applications (FLTA)
File in questo prodotto:
File Dimensione Formato  
A_Layer-Wise_Personalization_Approach_for_Transformer-Based_Federated_Anomaly_Detection.pdf

Accesso riservato

: Publisher’s version
Dimensione 564.73 kB
Formato Adobe PDF
564.73 kB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1281345
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 1
social impact