论文标题
对象识别的彩票票证假设
The Lottery Ticket Hypothesis for Object Recognition
论文作者
论文摘要
近年来,识别任务(例如对象识别和关键点估计)已广泛采用。这些任务的大多数最新方法都使用计算昂贵且具有巨大内存足迹的深网。这使得在低功率嵌入式设备上部署这些系统非常困难。因此,减少存储要求的重要性和此类模型中的计算量至关重要。最近提出的彩票票证假说(LTH)指出,在大型数据集中训练的深神经网络包含较小的子网,这些子网作为密集网络的par性能实现。在这项工作中,我们在对象检测,实例分割和关键点估计的背景下进行了第一个研究LTH的实证研究。我们的研究表明,从ImageNet进行预处理获得的彩票不能很好地转移到下游任务。我们提供有关如何在不同子任务上查找高达80%的总体稀疏性的彩票票的指导,而不会产生任何表现下降。最后,我们分析了受过训练的门票的行为相对于各种任务属性,例如对象大小,频率和检测难度。
Recognition tasks, such as object recognition and keypoint estimation, have seen widespread adoption in recent years. Most state-of-the-art methods for these tasks use deep networks that are computationally expensive and have huge memory footprints. This makes it exceedingly difficult to deploy these systems on low power embedded devices. Hence, the importance of decreasing the storage requirements and the amount of computation in such models is paramount. The recently proposed Lottery Ticket Hypothesis (LTH) states that deep neural networks trained on large datasets contain smaller subnetworks that achieve on par performance as the dense networks. In this work, we perform the first empirical study investigating LTH for model pruning in the context of object detection, instance segmentation, and keypoint estimation. Our studies reveal that lottery tickets obtained from ImageNet pretraining do not transfer well to the downstream tasks. We provide guidance on how to find lottery tickets with up to 80% overall sparsity on different sub-tasks without incurring any drop in the performance. Finally, we analyse the behavior of trained tickets with respect to various task attributes such as object size, frequency, and difficulty of detection.