Edge-Based Detection of Label Flipping Attacks in Federated Learning Using Explainable AI

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Federated Learning (FL) is a decentralized machine learning approach that enables collaborative training among distributed clients while preserving data privacy, making it increasingly popular for privacy-sensitive applications over traditional centralized models. However, it introduces new security vulnerabilities that challenge conventional approaches to software vulnerability management. Among these, label flipping attacks (LFAs) - where malicious clients intentionally mislabel data - pose a unique threat to the integrity of FL models. This study presents an AI-driven, edge-based vulnerability detection technique, leveraging explainable AI (XAI) techniques to enhance edge-based security within FL environments. Our method combines Grad-CAM visualizations with DBSCAN clustering to analyze class-specific behavior across clients. By detecting anomalies in Grad-CAM activation patterns, we identify malicious clients with flipped class labels, exploiting patterns in their Grad-CAM heatmaps. This approach is particularly robust against LFAs, examining each class independently and capturing patterns without relying on global model behavior. Empirical results on benchmark datasets such as MNIST and FashionMNIST demonstrate that our method accurately detects LFAs, even when malicious clients constitute a substantial portion of the network. This class-specific, XAI-driven approach contributes to the security of FL by offering an explainable, and scalable solution for managing vulnerabilities in distributed AI systems.

Original languageEnglish
Title of host publicationProceedings - 2025 IEEE/ACM 3rd International Workshop on Software Vulnerability Management, SVM 2025
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages1-8
Number of pages8
ISBN (Electronic)9798331514686
DOIs
StatePublished - 2025
Event3rd IEEE/ACM International Workshop on Software Vulnerability Management, SVM 2025 - Ottawa, Canada
Duration: 3 May 2025 → …

Publication series

NameProceedings - 2025 IEEE/ACM 3rd International Workshop on Software Vulnerability Management, SVM 2025

Conference

Conference3rd IEEE/ACM International Workshop on Software Vulnerability Management, SVM 2025
Country/TerritoryCanada
CityOttawa
Period3/05/25 → …

Bibliographical note

Publisher Copyright:
© 2025 IEEE.

Keywords

  • Grad-CAM
  • explainable AI
  • federated learning
  • label flipping attacks

ASJC Scopus subject areas

  • Software
  • Safety, Risk, Reliability and Quality

Fingerprint

Dive into the research topics of 'Edge-Based Detection of Label Flipping Attacks in Federated Learning Using Explainable AI'. Together they form a unique fingerprint.

Cite this