论文标题
VINNA:基于变异推理的神经网络体系结构搜索
VINNAS: Variational Inference-based Neural Network Architecture Search
论文作者
论文摘要
近年来,由于其能力在各种人工智能任务(例如图像分类或对象检测)中找到具有很高准确性的神经体系结构,因此神经建筑搜索(NAS)获得了密集的科学和工业兴趣。特别是,由于搜索过程中其计算效率,基于梯度的NAS方法已成为最受欢迎的方法之一。但是,这些方法通常会经历模式崩溃,在这种模式下,由于算法诉诸于为整个网络选择一种操作类型,或者在各种数据集或搜索空间中停滞在本地最小值,因此发现的体系结构的质量很差。 为了解决这些缺陷,我们提出了一种基于可区分的基于变异推理的NAS方法,用于搜索稀疏卷积神经网络。我们的方法通过使用自动相关性确定的各种辍学量在过度参数的超码中删除候选操作,从而找到了最佳的神经架构,这使该算法逐渐逐渐消除了不必要的操作和连接,而不会冒险崩溃。评估是通过搜索两种类型的卷积细胞来进行的,这些卷积细胞塑造了神经网络,用于分类不同的图像数据集。我们的方法找到了不同的网络单元,同时显示出最先进的准确性,其非零参数少了几乎2倍。
In recent years, neural architecture search (NAS) has received intensive scientific and industrial interest due to its capability of finding a neural architecture with high accuracy for various artificial intelligence tasks such as image classification or object detection. In particular, gradient-based NAS approaches have become one of the more popular approaches thanks to their computational efficiency during the search. However, these methods often experience a mode collapse, where the quality of the found architectures is poor due to the algorithm resorting to choosing a single operation type for the entire network, or stagnating at a local minima for various datasets or search spaces. To address these defects, we present a differentiable variational inference-based NAS method for searching sparse convolutional neural networks. Our approach finds the optimal neural architecture by dropping out candidate operations in an over-parameterised supergraph using variational dropout with automatic relevance determination prior, which makes the algorithm gradually remove unnecessary operations and connections without risking mode collapse. The evaluation is conducted through searching two types of convolutional cells that shape the neural network for classifying different image datasets. Our method finds diverse network cells, while showing state-of-the-art accuracy with up to almost 2 times fewer non-zero parameters.