Defense Strategies Toward Model Poisoning Attacks in Federated Learning: A Survey

Date
2022-04
Language
American English
Embargo Lift Date
Committee Members
Degree
Degree Year
Department
Grantor
Journal Title
Journal ISSN
Volume Title
Found At
IEEE
Abstract

Advances in distributed machine learning can empower future communications and networking. The emergence of federated learning (FL) has provided an efficient framework for distributed machine learning, which, however, still faces many security challenges. Among them, model poisoning attacks have a significant impact on the security and performance of FL. Given that there have been many studies focusing on defending against model poisoning attacks, it is necessary to survey the existing work and provide insights to inspire future research. In this paper, we first classify defense mechanisms for model poisoning attacks into two categories: evaluation methods for local model updates and aggregation methods for the global model. Then, we analyze some of the existing defense strategies in detail. We also discuss some potential challenges and future research directions. To the best of our knowledge, we are the first to survey defense methods for model poisoning attacks in FL.

Description
item.page.description.tableofcontents
item.page.relation.haspart
Cite As
Wang, Z., Kang, Q., Zhang, X., & Hu, Q. (2022). Defense Strategies Toward Model Poisoning Attacks in Federated Learning: A Survey. 2022 IEEE Wireless Communications and Networking Conference (WCNC), 548–553. https://doi.org/10.1109/WCNC51071.2022.9771619
ISSN
Publisher
Series/Report
Sponsorship
Major
Extent
Identifier
Relation
Journal
2022 IEEE Wireless Communications and Networking Conference (WCNC)
Rights
Publisher Policy
Source
Author
Alternative Title
Type
Conference proceedings
Number
Volume
Conference Dates
Conference Host
Conference Location
Conference Name
Conference Panel
Conference Secretariat Location
Version
Author's manuscript
Full Text Available at
This item is under embargo {{howLong}}