View a PDF of the paper titled Adversarial Robustness Unhardening via Backdoor Attacks in Federated Learning, by Taejin Kim and 4 other authors
View PDF
HTML (experimental)
Abstract:The delicate equilibrium between user privacy and the ability to unleash the potential of distributed data is an important concern. Federated learning, which enables the training of collaborative models without sharing of data, has emerged as a privacy-centric solution. This approach brings forth security challenges, notably poisoning and backdoor attacks where malicious entities inject corrupted data into the training process, as well as evasion attacks that aim to induce misclassifications at test time. Our research investigates the intersection of adversarial training, a common defense method against evasion attacks, and backdoor attacks within federated learning. We introduce Adversarial Robustness Unhardening (ARU), which is employed by a subset of adversarial clients to intentionally undermine model robustness during federated training, rendering models susceptible to a broader range of evasion attacks. We present extensive experiments evaluating ARU’s impact on adversarial training and existing robust aggregation defenses against poisoning and backdoor attacks. Our results show that ARU can substantially undermine adversarial training’s ability to harden models against test-time evasion attacks, and that adversaries employing ARU can even evade robust aggregation defenses that often neutralize poisoning or backdoor attacks.
Submission history
From: Taejin Kim [view email]
[v1]
Tue, 17 Oct 2023 21:38:41 UTC (663 KB)
[v2]
Sat, 21 Oct 2023 03:18:35 UTC (663 KB)
[v3]
Sun, 29 Jun 2025 19:25:01 UTC (2,189 KB)