论文标题

FEDSL:在复发性神经网络中的分布式顺序数据上联合拆分学习

FedSL: Federated Split Learning on Distributed Sequential Data in Recurrent Neural Networks

论文作者

Abedi, Ali, Khan, Shehroz S.

论文摘要

联合学习(FL)和拆分学习(SL)是隐私的机器学习(ML)技术,可以使培训ML模型超过客户之间的数据,而无需直接访问其原始数据。现有的FL和SL方法在水平或垂直分区的数据上工作,并且无法处理依次分区的数据,在这些数据中,多段顺序数据的段分布在客户端之间。在本文中,我们提出了一个新颖的联合分裂学习框架FEDSL,以在分布式顺序数据上训练模型。序列数据训练的最常见的ML模型是复发性神经网络(RNN)。由于所提出的框架是隐私保护的,因此在客户端或客户端和服务器之间无法共享多段顺序数据的段。为了规避这一限制,我们提出了一种针对RNN量身定制的新型SL方法。 RNN被分为子网络,每个子网络都经过一个访问的一个客户端,该客户端包含一个多段训练序列的单个段。在本地培训期间,不同客户端的子网络相互通信,以捕获不同客户端多段顺序数据的连续段之间的潜在依赖项,但没有共享原始数据或完整的模型参数。在使用本地顺序数据段培训本地子网络之后,所有客户端将其子网络发送到联合服务器,该服务器被汇总为生成全局模型。对模拟和现实世界数据集的实验结果表明,所提出的方法成功地在分布式的顺序数据上训练模型,同时保留隐私,并且在更少的沟通回合中实现更高准确性的方面优于先前的FL和集中学习方法。

Federated Learning (FL) and Split Learning (SL) are privacy-preserving Machine-Learning (ML) techniques that enable training ML models over data distributed among clients without requiring direct access to their raw data. Existing FL and SL approaches work on horizontally or vertically partitioned data and cannot handle sequentially partitioned data where segments of multiple-segment sequential data are distributed across clients. In this paper, we propose a novel federated split learning framework, FedSL, to train models on distributed sequential data. The most common ML models to train on sequential data are Recurrent Neural Networks (RNNs). Since the proposed framework is privacy-preserving, segments of multiple-segment sequential data cannot be shared between clients or between clients and server. To circumvent this limitation, we propose a novel SL approach tailored for RNNs. A RNN is split into sub-networks, and each sub-network is trained on one client containing single segments of multiple-segment training sequences. During local training, the sub-networks on different clients communicate with each other to capture latent dependencies between consecutive segments of multiple-segment sequential data on different clients, but without sharing raw data or complete model parameters. After training local sub-networks with local sequential data segments, all clients send their sub-networks to a federated server where sub-networks are aggregated to generate a global model. The experimental results on simulated and real-world datasets demonstrate that the proposed method successfully trains models on distributed sequential data, while preserving privacy, and outperforms previous FL and centralized learning approaches in terms of achieving higher accuracy in fewer communication rounds.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源