Li Ruiheng, Cao Yue, Shu Yuhang, Guo Jia, Shi Binghua, Yu Jiaojiao, Di Yi, Zuo Qiankun, Tian Hao
Hubei Key Laboratory of Digital Finance Innovation, Hubei University of Economics, Wuhan, 430205, China.
School of Information Engineering, Hubei University of Economics, Wuhan, 430205, China.
Sci Rep. 2024 Nov 3;14(1):26515. doi: 10.1038/s41598-024-77310-z.
Federated Learning (FL) uses local data to perform distributed training on clients and combines resulting models on a public server to mitigate privacy exposure by avoiding data sharing. However, further research indicates that communication overheads continue to be the primary limitation for FL relative to alternative considerations. This is especially true when training models on non-independent and identically distributed data, such as financial default risk data, where FL's computational costs increase significantly. This study aims to address financial credit risk data by establishing a dynamic receptive field adjustment mechanism for feature extraction, efficiently extracting default features with varying distributions and attributes. Additionally, by constructing a distributed feature fusion architecture, characteristics from both local and overarching models are aggregated to attain higher accuracy with lower transmission costs. Experimental results demonstrate that the proposed FL framework can utilize dynamic receptive fields to allocate convolutional kernel weights, thereby improving feature extraction. In the feature fusion stage, the proposed Multi-Fusion strategy efficiently customizes the aggregation of features from local and global models. The final solution reduces the communication rounds in federated learning by approximately 80%.
联邦学习(FL)利用本地数据在客户端进行分布式训练,并在公共服务器上合并生成的模型,以通过避免数据共享来减轻隐私暴露。然而,进一步的研究表明,相对于其他考虑因素,通信开销仍然是联邦学习的主要限制。在对非独立同分布数据(如金融违约风险数据)进行模型训练时尤其如此,此时联邦学习的计算成本会显著增加。本研究旨在通过建立用于特征提取的动态感受野调整机制来处理金融信用风险数据,有效提取具有不同分布和属性的违约特征。此外,通过构建分布式特征融合架构,聚合本地模型和总体模型的特征,以更低的传输成本实现更高的准确性。实验结果表明,所提出的联邦学习框架可以利用动态感受野来分配卷积核权重,从而改进特征提取。在特征融合阶段,所提出的多融合策略有效地定制了本地模型和全局模型特征的聚合。最终方案将联邦学习中的通信轮次减少了约80%。