论文标题
与嘈杂的标签的强大联盟学习
Robust Federated Learning with Noisy Labels
论文作者
论文摘要
联合学习是一种范式,它使本地设备能够共同训练服务器模型,同时保持数据分散和私密。在联合学习中,由于客户收集了本地数据,因此无法保证数据正确注释。尽管已经进行了许多研究,以在集中式的环境中培训这些嘈杂数据的网络,但这些算法在联邦学习中仍然具有嘈杂的标签。与集中设置相比,由于其标签系统的变化或用户的背景知识,客户的数据可能具有不同的噪声分布。结果,本地模型构成了不一致的决策界限及其权重,彼此之间存在严重分歧,这在联邦学习中是严重的问题。为了解决这些问题,我们介绍了一种新颖的联合学习计划,服务器与本地模型合作,以通过互换班级质心来保持一致的决策边界。这些质心是每个设备上本地数据的主要特征,每个设备都由服务器对齐每个通信。使用对齐的质心更新本地模型有助于在本地模型之间形成一致的决策边界,尽管客户数据中的噪声分布彼此不同。为了提高本地模型性能,我们引入了一种新颖的方法,以选择用于更新给定标签模型的自信样本。此外,我们提出了一种全球引导的伪标记方法,通过利用全球模型来更新非义务样本的标签。我们对嘈杂的CIFAR-10数据集和Clothing1M数据集的实验结果表明,我们的方法在使用嘈杂的标签的联合学习中明显有效。
Federated learning is a paradigm that enables local devices to jointly train a server model while keeping the data decentralized and private. In federated learning, since local data are collected by clients, it is hardly guaranteed that the data are correctly annotated. Although a lot of studies have been conducted to train the networks robust to these noisy data in a centralized setting, these algorithms still suffer from noisy labels in federated learning. Compared to the centralized setting, clients' data can have different noise distributions due to variations in their labeling systems or background knowledge of users. As a result, local models form inconsistent decision boundaries and their weights severely diverge from each other, which are serious problems in federated learning. To solve these problems, we introduce a novel federated learning scheme that the server cooperates with local models to maintain consistent decision boundaries by interchanging class-wise centroids. These centroids are central features of local data on each device, which are aligned by the server every communication round. Updating local models with the aligned centroids helps to form consistent decision boundaries among local models, although the noise distributions in clients' data are different from each other. To improve local model performance, we introduce a novel approach to select confident samples that are used for updating the model with given labels. Furthermore, we propose a global-guided pseudo-labeling method to update labels of unconfident samples by exploiting the global model. Our experimental results on the noisy CIFAR-10 dataset and the Clothing1M dataset show that our approach is noticeably effective in federated learning with noisy labels.