论文标题

在深厚的加强学习任务中,机器与人类的关注

Machine versus Human Attention in Deep Reinforcement Learning Tasks

论文作者

Guo, Sihang, Zhang, Ruohan, Liu, Bo, Zhu, Yifeng, Hayhoe, Mary, Ballard, Dana, Stone, Peter

论文摘要

深度强化学习(RL)算法是解决视觉运动决策任务的强大工具。但是,训练有素的模型通常很难解释,因为它们被表示为端到端的深神经网络。在本文中,我们通过分析任务执行期间所使用的像素,并将它们与人类执行相同任务的像素进行比较,从而阐明了这些受过训练的模型的内部工作。为此,我们研究了以下两个问题,据我们所知,以前尚未研究过。 1)在执行相同任务时,RL代理和人类学到的视觉表示如何相似? 2)这些学识渊博的表示形式中的相似性和差异如何解释RL代理在这些任务上的表现?具体而言,我们将RL代理的显着性图与学习玩Atari游戏时的人类专家的视觉关注模型进行了比较。此外,我们分析了深度RL算法的超参数如何影响训练有素的代理的学会表示和显着图。提供的见解有可能告知新型算法,以缩小人类专家与RL代理之间的性能差距。

Deep reinforcement learning (RL) algorithms are powerful tools for solving visuomotor decision tasks. However, the trained models are often difficult to interpret, because they are represented as end-to-end deep neural networks. In this paper, we shed light on the inner workings of such trained models by analyzing the pixels that they attend to during task execution, and comparing them with the pixels attended to by humans executing the same tasks. To this end, we investigate the following two questions that, to the best of our knowledge, have not been previously studied. 1) How similar are the visual representations learned by RL agents and humans when performing the same task? and, 2) How do similarities and differences in these learned representations explain RL agents' performance on these tasks? Specifically, we compare the saliency maps of RL agents against visual attention models of human experts when learning to play Atari games. Further, we analyze how hyperparameters of the deep RL algorithm affect the learned representations and saliency maps of the trained agents. The insights provided have the potential to inform novel algorithms for closing the performance gap between human experts and RL agents.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源