论文标题
对抗性强大的神经体系结构
Adversarially Robust Neural Architectures
论文作者
论文摘要
深度神经网络(DNN)容易受到对抗攻击的影响。现有方法致力于制定各种强大的培训策略或正规化,以更新神经网络的权重。但是,除了权重之外,网络中的总体结构和信息流由神经体系结构明确确定,而神经结构仍未得到探索。因此,本文旨在从体系结构的角度提高网络的对抗性鲁棒性。我们探讨了对抗性鲁棒性,Lipschitz常数和体系结构参数之间的关系,并表明对体系结构参数的适当约束可以减少Lipschitz常数以进一步提高鲁棒性。体系结构参数的重要性可能因操作到操作或连接连接而异。我们通过单变量对数正态分布近似整个网络的Lipschitz常数,其平均值和方差与架构参数有关。可以通过基于累积函数对分布参数制定约束来实现置信度。与通过各种NAS算法搜索的对抗训练的神经体系结构以及有效的人为设计的模型相比,我们的算法在各种攻击下在不同数据集中的各种攻击下实现了最佳性能。
Deep Neural Networks (DNNs) are vulnerable to adversarial attacks. Existing methods are devoted to developing various robust training strategies or regularizations to update the weights of the neural network. But beyond the weights, the overall structure and information flow in the network are explicitly determined by the neural architecture, which remains unexplored. This paper thus aims to improve the adversarial robustness of the network from the architecture perspective. We explore the relationship among adversarial robustness, Lipschitz constant, and architecture parameters and show that an appropriate constraint on architecture parameters could reduce the Lipschitz constant to further improve the robustness. The importance of architecture parameters could vary from operation to operation or connection to connection. We approximate the Lipschitz constant of the entire network through a univariate log-normal distribution, whose mean and variance are related to architecture parameters. The confidence can be fulfilled through formulating a constraint on the distribution parameters based on the cumulative function. Compared with adversarially trained neural architectures searched by various NAS algorithms as well as efficient human-designed models, our algorithm empirically achieves the best performance among all the models under various attacks on different datasets.