论文标题
用于人工神经网络的强大培训的分析框架
An Analytic Framework for Robust Training of Artificial Neural Networks
论文作者
论文摘要
学习模型的可靠性是成功部署机器学习在各个行业的关键。创建一个健壮的模型,尤其是一个不受对抗攻击影响的模型,需要对对抗性实例的现象有全面的了解。但是,由于机器学习中问题的复杂性,很难描述这种现象。因此,许多研究通过提出一个简化的模型来研究这种现象,该模型是如何发生对抗性例子并通过预测现象的某些方面进行验证。尽管这些研究涵盖了对抗性实例的许多不同特征,但它们尚未对现象的几何和分析建模采用整体方法。本文提出了一个正式的框架,以研究学习理论中的现象,并利用复杂的分析和全体形态,为人工神经网络提供强大的学习规则。借助复杂的分析,我们可以在现象的几何和分析观点之间毫不费力地移动,并通过揭示其与谐波功能的联系来进一步就该现象的见解。使用我们的模型,我们可以解释对抗性例子的一些最有趣的特征,包括对抗性实例的可转移性,并为减轻现象的效果的新方法铺平了道路。
The reliability of a learning model is key to the successful deployment of machine learning in various industries. Creating a robust model, particularly one unaffected by adversarial attacks, requires a comprehensive understanding of the adversarial examples phenomenon. However, it is difficult to describe the phenomenon due to the complicated nature of the problems in machine learning. Consequently, many studies investigate the phenomenon by proposing a simplified model of how adversarial examples occur and validate it by predicting some aspect of the phenomenon. While these studies cover many different characteristics of the adversarial examples, they have not reached a holistic approach to the geometric and analytic modeling of the phenomenon. This paper propose a formal framework to study the phenomenon in learning theory and make use of complex analysis and holomorphicity to offer a robust learning rule for artificial neural networks. With the help of complex analysis, we can effortlessly move between geometric and analytic perspectives of the phenomenon and offer further insights on the phenomenon by revealing its connection with harmonic functions. Using our model, we can explain some of the most intriguing characteristics of adversarial examples, including transferability of adversarial examples, and pave the way for novel approaches to mitigate the effects of the phenomenon.