论文标题
解决联邦学习中的班级失衡
Addressing Class Imbalance in Federated Learning
论文作者
论文摘要
联合学习(FL)是一种有前途的方法,用于培训位于本地客户设备上的分散数据,同时提高效率和隐私。但是,对客户方面的培训数据的分布和数量可能会带来重大挑战,例如类别不平衡和非IID(非独立且相同分布的数据),这可能会极大地影响通用模型的性能。尽管在遇到非IID数据时已大量的努力致力于帮助FL模型融合,但不平衡问题尚未得到充分解决。特别是,由于FL培训是通过以加密形式交换梯度来执行的,因此培训数据对客户或服务器都无法完全观察到,并且以前的类不平衡方法对FL的表现不佳。因此,设计用于检测FL中类不平衡并减轻其影响的新方法至关重要。在这项工作中,我们提出了一个监视方案,可以推断每个FL回合的训练数据组成,并设计一个新的损失函数 - \ textbf {比率损失},以减轻失衡的影响。我们的实验表明,在FL培训中承认阶级失衡并尽早采取措施的重要性,以及我们方法在减轻影响方面的有效性。我们的方法显示在维护客户隐私的同时,可以显着胜过以前的方法。
Federated learning (FL) is a promising approach for training decentralized data located on local client devices while improving efficiency and privacy. However, the distribution and quantity of the training data on the clients' side may lead to significant challenges such as class imbalance and non-IID (non-independent and identically distributed) data, which could greatly impact the performance of the common model. While much effort has been devoted to helping FL models converge when encountering non-IID data, the imbalance issue has not been sufficiently addressed. In particular, as FL training is executed by exchanging gradients in an encrypted form, the training data is not completely observable to either clients or servers, and previous methods for class imbalance do not perform well for FL. Therefore, it is crucial to design new methods for detecting class imbalance in FL and mitigating its impact. In this work, we propose a monitoring scheme that can infer the composition of training data for each FL round, and design a new loss function -- \textbf{Ratio Loss} to mitigate the impact of the imbalance. Our experiments demonstrate the importance of acknowledging class imbalance and taking measures as early as possible in FL training, and the effectiveness of our method in mitigating the impact. Our method is shown to significantly outperform previous methods, while maintaining client privacy.