论文标题

通过自我监督实例过滤和错误映射启用设备CNN培训

Enabling On-Device CNN Training by Self-Supervised Instance Filtering and Error Map Pruning

论文作者

Wu, Yawen, Wang, Zhepeng, Shi, Yiyu, Hu, Jingtong

论文摘要

这项工作旨在通过降低培训时的计算成本来实现卷积神经网络(CNN)的设备培训。 CNN型号通常在高性能计算机上进行训练,并且仅将训练有素的型号部署到边缘设备。但是,经过静态训练的模型不能在真实环境中动态调整,并且可能导致新输入的精度较低。部署后通过从现实世界数据中学习来大大提高准确性的培训。但是,高计算成本使得对资源受限设备的培训过高。为了解决此问题,我们通过两种互补方法探索培训中的计算冗余,并降低计算成本:在数据级别上进行自我监督的早期实例过滤,并在算法级别上修剪错误。早期实例过滤器从输入流选择重要实例以训练网络并删除琐碎的实例。在使用选定实例训练时,错误映射会进一步修剪不重要的计算。广泛的实验表明,计算成本大大降低,而没有任何或边际准确性损失。例如,在CIFAR-10上训练RESNET-11时,我们可以节省68%的计算,同时保持准确性和75%的计算节省,而边缘精度损失为1.3%。当量化整合到建议的方法中时,积极的计算节省96%的精度损失少于0.1%。此外,在MNIST上训练LENET时,我们节省了79%的计算,同时将精度提高了0.2%。

This work aims to enable on-device training of convolutional neural networks (CNNs) by reducing the computation cost at training time. CNN models are usually trained on high-performance computers and only the trained models are deployed to edge devices. But the statically trained model cannot adapt dynamically in a real environment and may result in low accuracy for new inputs. On-device training by learning from the real-world data after deployment can greatly improve accuracy. However, the high computation cost makes training prohibitive for resource-constrained devices. To tackle this problem, we explore the computational redundancies in training and reduce the computation cost by two complementary approaches: self-supervised early instance filtering on data level and error map pruning on the algorithm level. The early instance filter selects important instances from the input stream to train the network and drops trivial ones. The error map pruning further prunes out insignificant computations when training with the selected instances. Extensive experiments show that the computation cost is substantially reduced without any or with marginal accuracy loss. For example, when training ResNet-110 on CIFAR-10, we achieve 68% computation saving while preserving full accuracy and 75% computation saving with a marginal accuracy loss of 1.3%. Aggressive computation saving of 96% is achieved with less than 0.1% accuracy loss when quantization is integrated into the proposed approaches. Besides, when training LeNet on MNIST, we save 79% computation while boosting accuracy by 0.2%.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源