论文标题
使用捕获范围的挑战对渗透测试进行建模:通过捕获挑战:无模型学习与先验知识之间的权衡
Modeling Penetration Testing with Reinforcement Learning Using Capture-the-Flag Challenges: Trade-offs between Model-free Learning and A Priori Knowledge
论文作者
论文摘要
渗透测试是一种旨在通过模拟对其进行攻击来评估系统安全性的安全练习。到目前为止,渗透测试主要是由训练有素的人类攻击者进行的,其成功取决于可用的专业知识。自动化这种做法构成了一个非平凡的问题,因为人类专家可能试图抵制系统的一系列行动以及她所依赖的知识范围很难捕获。在本文中,我们将注意力集中在简化的渗透测试问题上以捕获旗帜黑客入侵挑战的形式表达的问题,并分析了无模型的增强学习算法如何帮助解决方案。在对这些捕获的旗帜竞争作为强化学习问题时,我们强调,表征渗透测试的特定挑战是发现手头问题的结构的问题。然后,我们通过依靠可能提供给代理商的不同形式的先验知识来显示这一挑战。通过这种方式,我们演示了使用增强学习来应对渗透测试的可行性如何取决于无模型和基于模型的算法之间的仔细权衡。通过使用技术注入先验知识,我们表明可以更好地指导代理并限制其勘探问题的空间,从而更有效地实现解决方案。
Penetration testing is a security exercise aimed at assessing the security of a system by simulating attacks against it. So far, penetration testing has been carried out mainly by trained human attackers and its success critically depended on the available expertise. Automating this practice constitutes a non-trivial problem, as the range of actions that a human expert may attempts against a system and the range of knowledge she relies on to take her decisions are hard to capture. In this paper, we focus our attention on simplified penetration testing problems expressed in the form of capture the flag hacking challenges, and we analyze how model-free reinforcement learning algorithms may help to solve them. In modeling these capture the flag competitions as reinforcement learning problems we highlight that a specific challenge that characterize penetration testing is the problem of discovering the structure of the problem at hand. We then show how this challenge may be eased by relying on different forms of prior knowledge that may be provided to the agent. In this way we demonstrate how the feasibility of tackling penetration testing using reinforcement learning may rest on a careful trade-off between model-free and model-based algorithms. By using techniques to inject a priori knowledge, we show it is possible to better direct the agent and restrict the space of its exploration problem, thus achieving solutions more efficiently.