论文标题

监督鲁棒性的无数据神经网络修剪

Supervised Robustness-preserving Data-free Neural Network Pruning

论文作者

Meng, Mark Huasong, Bai, Guangdong, Teo, Sin Gee, Dong, Jin Song

论文摘要

在现实世界应用程序中部署预训练的神经网络模型时,模型消费者通常会遇到资源构成平台,例如移动设备和智能设备。他们通常使用修剪技术来降低模型的尺寸和复杂性,从而产生更轻的资源消耗量。尽管如此,大多数现有的修剪方法都是在修剪后的模型有机会根据原始培训数据进行微调甚至重新训练的前提。实际上,这可能是不现实的,因为数据控制器通常不愿为其模型消费者提供原始数据。在这项工作中,我们研究了无数据上下文中的神经网络修剪,旨在产生轻巧的模型,这些模型不仅在预测中准确,而且对开放世界部署中不希望的输入也有强大的稳定性。考虑到缺乏可以修复错误的单元的微调和再培训,我们用保守的单元替换传统的侵略性单枪策略,将修剪作为渐进式过程。我们提出了一种基于随机优化的修剪方法,该方法使用鲁棒性相关的指标来指导修剪过程。我们的方法是作为Python程序实施的,并通过一系列有关不同神经网络模型的实验进行了评估。实验结果表明,就稳健性和准确性而言,它的表现明显优于现有的无数据修剪方法。

When deploying pre-trained neural network models in real-world applications, model consumers often encounter resource-constraint platforms such as mobile and smart devices. They typically use the pruning technique to reduce the size and complexity of the model, generating a lighter one with less resource consumption. Nonetheless, most existing pruning methods are proposed with the premise that the model after being pruned has a chance to be fine-tuned or even retrained based on the original training data. This may be unrealistic in practice, as the data controllers are often reluctant to provide their model consumers with the original data. In this work, we study the neural network pruning in the data-free context, aiming to yield lightweight models that are not only accurate in prediction but also robust against undesired inputs in open-world deployments. Considering the absence of the fine-tuning and retraining that can fix the mis-pruned units, we replace the traditional aggressive one-shot strategy with a conservative one that treats the pruning as a progressive process. We propose a pruning method based on stochastic optimization that uses robustness-related metrics to guide the pruning process. Our method is implemented as a Python program and evaluated with a series of experiments on diverse neural network models. The experimental results show that it significantly outperforms existing one-shot data-free pruning approaches in terms of robustness preservation and accuracy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源