论文标题
联合课堂学习
Federated Class-Incremental Learning
论文作者
论文摘要
联邦学习(FL)通过对分散客户的数据私有协作培训吸引了日益增长的关注。但是,大多数现有的方法不切实际地假设整个框架的对象类是随着时间的推移而固定的。它使全球模型在现实情况下遭受了旧课程的巨大灾难性遗忘,在那里,本地客户经常连续收集新课程,并且存储旧课程的存储内存非常有限。此外,拥有新班的新客户可能会参加FL培训,从而进一步加剧了对全球模型的灾难性遗忘。为了应对这些挑战,我们开发了一种新型的全球本地遗忘薪酬(GLFC)模型,以学习一种从本地和全球观点来减轻灾难性遗忘的全球班级增量模型。具体来说,为了解决当地客户的班级失衡引起的本地遗忘,我们设计了班级感知的梯度薪酬损失和班级语义关系蒸馏损失,以平衡遗忘旧班级和跨任务跨越阶层的一致性关系。为了解决非I.I.D类不平衡跨客户的全球遗忘,我们提出了一个代理服务器,该服务器选择了最佳的旧全球模型来协助本地关系蒸馏。此外,开发了一种基于原型梯度的通信机制来保护隐私。就代表性基准数据集的平均准确性而言,我们的模型优于最先进的方法的最先进方法。
Federated learning (FL) has attracted growing attention via data-private collaborative training on decentralized clients. However, most existing methods unrealistically assume object classes of the overall framework are fixed over time. It makes the global model suffer from significant catastrophic forgetting on old classes in real-world scenarios, where local clients often collect new classes continuously and have very limited storage memory to store old classes. Moreover, new clients with unseen new classes may participate in the FL training, further aggravating the catastrophic forgetting of the global model. To address these challenges, we develop a novel Global-Local Forgetting Compensation (GLFC) model, to learn a global class incremental model for alleviating the catastrophic forgetting from both local and global perspectives. Specifically, to address local forgetting caused by class imbalance at the local clients, we design a class-aware gradient compensation loss and a class-semantic relation distillation loss to balance the forgetting of old classes and distill consistent inter-class relations across tasks. To tackle the global forgetting brought by the non-i.i.d class imbalance across clients, we propose a proxy server that selects the best old global model to assist the local relation distillation. Moreover, a prototype gradient-based communication mechanism is developed to protect privacy. Our model outperforms state-of-the-art methods by 4.4%-15.1% in terms of average accuracy on representative benchmark datasets.