A Dual-Defense Self-balancing Framework Against Bilateral Model Attacks in Federated Learning

Xiang WU, Aiting YAO, Shantanu PAL, Frank JIANG, Xuejun LI*, Jia XU, Chengzu DONG, Xuefei CHEN, Xiuyi ZHANG, Xiao LIU

*Corresponding author for this work

Research output: Book Chapters | Papers in Conference ProceedingsConference paper (refereed)Researchpeer-review

Abstract

With the rapid expansion of Artificial Intelligence (AI) services, smart devices generate a large amount of user data at the edge network, which urgently needs to be protected while effectively extracting information. Federated learning (FL) is an important technology for handling dispersed data and strict privacy requirements in this context. However, the security threats caused by model inversion attacks and poisoning attacks can affect the mutual trust between the client and server. Yet, for these two types of attacks, the existing defense mechanisms are contradictory in terms of whether the model parameters are publicly disclosed. In addition, the data distribution of the clients is imbalanced which will increase the bias of model, reducing its practicality. To address this issue, this study proposes a dual defense self-balanced federated learning (DDSFL) framework, aiming to introduce a novel lightweight defense mechanism during the model parameter aggregation stage, combating these two types of attacks simultaneously by applying differential privacy and adjusting learning rates. In addition, this method also integrates a middleware-based reordering algorithm to enhance the robustness of the framework. Experimental results show that DDSFL effectively improves the ability to resist imbalanced data, forged data, and malicious behavior, significantly enhancing the generalization performance and security of the FL system.

Original languageEnglish
Title of host publicationAlgorithms and Architectures for Parallel Processing: 24th International Conference, ICA3PP 2024, Proceedings, Part I
EditorsTianqing ZHU, Jin LI, Aniello CASTIGLIONE
PublisherSpringer Science and Business Media Deutschland GmbH
Pages261-270
Number of pages10
ISBN (Print)9789819615247
DOIs
Publication statusPublished - 2025
Event24th International Conference on Algorithms and Architectures for Parallel Processing, ICA3PP 2024 - Macau, China
Duration: 29 Oct 202431 Oct 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume15251 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349
NameInternational Conference on Algorithms and Architectures for Parallel Processing
PublisherSpringer
VolumeICA3PP 2024

Conference

Conference24th International Conference on Algorithms and Architectures for Parallel Processing, ICA3PP 2024
Country/TerritoryChina
CityMacau
Period29/10/2431/10/24

Bibliographical note

Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.

Funding

This work was supported by the National Natural Science Foundation of China Project (No. 62372004).

Keywords

  • Differential Privacy
  • Edge Network
  • Federated Learning
  • Model Inversion Attacks
  • Model Poisoning Attacks

Fingerprint

Dive into the research topics of 'A Dual-Defense Self-balancing Framework Against Bilateral Model Attacks in Federated Learning'. Together they form a unique fingerprint.

Cite this