Tan Zhou, Cai Jianping, Li De, Lian Puwei, Liu Ximeng, Che Yan
College of Computer Science and Big Data, Fuzhou University, Fuzhou, 350000, China.
School of Computer Science and Engineering, Guangxi Normal University, Guilin, 541004, China.
Neural Netw. 2025 Apr;184:107016. doi: 10.1016/j.neunet.2024.107016. Epub 2024 Dec 10.
Federated Learning (FL) is an efficient, distributed machine learning paradigm that enables multiple clients to jointly train high-performance deep learning models while maintaining training data locally. However, due to its distributed computing nature, malicious clients can manipulate the prediction of the trained model through backdoor attacks. Existing defense methods require significant computational and communication overhead during the training or testing phases, limiting their practicality in resource-constrained scenarios and being unsuitable for the Non-IID data distribution typical in general FL scenarios. To address these challenges, we propose the FedPD framework, in which servers and clients exchange prototypes rather than model parameters, preventing the implantation of backdoor channels by malicious clients during FL training and effectively eliminating the success of backdoor attacks at the source, significantly reducing communication overhead. Additionally, prototypes can serve as global knowledge to correct clients' local training. Experiments and performance analysis show that FedPD achieves superior and consistent defense performance compared to existing representative approaches against backdoor attacks. In specific scenarios, FedPD can reduce the success rate of attacks by 90.73% compared to FedAvg without defense while maintaining the main task accuracy above 90%.
联邦学习(FL)是一种高效的分布式机器学习范式,它使多个客户端能够在本地保留训练数据的同时联合训练高性能深度学习模型。然而,由于其分布式计算的特性,恶意客户端可以通过后门攻击操纵训练模型的预测。现有的防御方法在训练或测试阶段需要大量的计算和通信开销,这限制了它们在资源受限场景中的实用性,并且不适用于一般联邦学习场景中典型的非独立同分布数据分布。为应对这些挑战,我们提出了FedPD框架,其中服务器和客户端交换原型而非模型参数,防止恶意客户端在联邦学习训练期间植入后门通道,并从源头上有效消除后门攻击的成功性,显著减少通信开销。此外,原型可以作为全局知识来纠正客户端的本地训练。实验和性能分析表明,与现有的针对后门攻击的代表性方法相比,FedPD实现了卓越且一致的防御性能。在特定场景中,与无防御的联邦平均算法(FedAvg)相比,FedPD可以将攻击成功率降低90.73%,同时将主要任务的准确率保持在90%以上。