论文标题
神经结构转移
Neural Architecture Transfer
论文作者
论文摘要
神经体系结构搜索(NAS)已成为自动设计特定于任务的神经网络的有前途的途径。现有的NAS方法需要对硬件或目标的每个部署规范进行全面搜索。鉴于潜在的应用程序场景,这是计算上不切实际的努力。在本文中,我们提出了神经结构转移(NAT)来克服这一限制。 NAT旨在有效地生成特定于任务的自定义模型,这些模型在多个冲突目标下具有竞争力。为了实现这一目标,我们学习了特定于任务的超级网,可以从中可以从中采样专门的子网,而无需任何其他培训。我们方法的关键是集成的在线转移学习和多个目标进化搜索程序。预先训练的超网是迭代的,同时搜索特定于任务的子网。我们证明了NAT对11个基准图像分类任务的功效,从大型多级到小规模的细粒数据集等等。在所有情况下,包括Imagenet在内的NATNET在移动设置下的最新设置($ \ leq $ 6亿美元的多重添加)都可以改善。出乎意料的是,小规模的细粒数据集中受益最大。同时,体系结构搜索和传输是比现有NAS方法更有效的数量级。总体而言,实验评估表明,在不同的图像分类任务和计算目标中,NAT是一种更有效的替代方法,可以替代传统转移学习在标准数据集中学到的现有网络体系结构的微调权重。代码可从https://github.com/human-analysis/neural-architecture-transfer获得
Neural architecture search (NAS) has emerged as a promising avenue for automatically designing task-specific neural networks. Existing NAS approaches require one complete search for each deployment specification of hardware or objective. This is a computationally impractical endeavor given the potentially large number of application scenarios. In this paper, we propose Neural Architecture Transfer (NAT) to overcome this limitation. NAT is designed to efficiently generate task-specific custom models that are competitive under multiple conflicting objectives. To realize this goal we learn task-specific supernets from which specialized subnets can be sampled without any additional training. The key to our approach is an integrated online transfer learning and many-objective evolutionary search procedure. A pre-trained supernet is iteratively adapted while simultaneously searching for task-specific subnets. We demonstrate the efficacy of NAT on 11 benchmark image classification tasks ranging from large-scale multi-class to small-scale fine-grained datasets. In all cases, including ImageNet, NATNets improve upon the state-of-the-art under mobile settings ($\leq$ 600M Multiply-Adds). Surprisingly, small-scale fine-grained datasets benefit the most from NAT. At the same time, the architecture search and transfer is orders of magnitude more efficient than existing NAS methods. Overall, the experimental evaluation indicates that, across diverse image classification tasks and computational objectives, NAT is an appreciably more effective alternative to conventional transfer learning of fine-tuning weights of an existing network architecture learned on standard datasets. Code is available at https://github.com/human-analysis/neural-architecture-transfer