论文标题
通过过滤修剪和知识转移在边缘设备上深入学习
Enabling Deep Learning on Edge Devices through Filter Pruning and Knowledge Transfer
论文作者
论文摘要
深度学习模型已将各种智能应用程序引入边缘设备,例如图像分类,语音识别和增强现实。为了提供个性化,响应式和私人学习,越来越需要在设备上培训此类模型。为了满足这一需求,本文提出了一种新的解决方案,用于在资源约束设备上部署和培训最先进的模型。首先,本文提出了一种基于新颖的基于滤波器的模型压缩方法,以从云中训练的大型模型中创建可轻巧的可训练模型,而没有太多的准确性损失。其次,它提出了一种新颖的知识转移方法,以使设备模型使用新数据上的增量学习实时或接近实时更新,并使设备模型能够以无聊的方式借助于露出的模型来学习看不见的类别。结果表明,1)我们的模型压缩方法最多可以删除WRN-28-10的99.36%参数,同时在CIFAR-10上保留超过90%的前1个精度; 2)我们的知识转移方法使压缩模型能够在CIFAR-10上实现超过90%的精度,并在旧类别上保持良好的精度; 3)它允许压缩模型在边缘的实时(三到六分钟)内收敛,以进行增量学习任务; 4)它使该模型能够对其从未经过训练的数据类别(78.92%的TOP-1准确性)进行分类。
Deep learning models have introduced various intelligent applications to edge devices, such as image classification, speech recognition, and augmented reality. There is an increasing need of training such models on the devices in order to deliver personalized, responsive, and private learning. To address this need, this paper presents a new solution for deploying and training state-of-the-art models on the resource-constrained devices. First, the paper proposes a novel filter-pruning-based model compression method to create lightweight trainable models from large models trained in the cloud, without much loss of accuracy. Second, it proposes a novel knowledge transfer method to enable the on-device model to update incrementally in real time or near real time using incremental learning on new data and enable the on-device model to learn the unseen categories with the help of the in-cloud model in an unsupervised fashion. The results show that 1) our model compression method can remove up to 99.36% parameters of WRN-28-10, while preserving a Top-1 accuracy of over 90% on CIFAR-10; 2) our knowledge transfer method enables the compressed models to achieve more than 90% accuracy on CIFAR-10 and retain good accuracy on old categories; 3) it allows the compressed models to converge within real time (three to six minutes) on the edge for incremental learning tasks; 4) it enables the model to classify unseen categories of data (78.92% Top-1 accuracy) that it is never trained with.