论文标题
部分可观测时空混沌系统的无模型预测
Reasoning about Counterfactuals to Improve Human Inverse Reinforcement Learning
论文作者
论文摘要
为了与机器人合作,我们必须能够理解他们的决策。人类自然会通过对其可观察到的行为进行推理,以类似于逆增强学习(IRL)的方式来推论其他代理人的信念和欲望。因此,机器人可以通过提供对人类学习者的IRL提供信息的示威来传达他们的信念和欲望。一项信息的演示是,鉴于他们当前对机器人的决策做出的理解,这一演示与学习者对机器人将要做的事情的期望有很大差异。但是,标准IRL并未对学习者的现有期望进行建模,因此不能执行这种反事实推理。我们建议将学习者对机器人决策制定的当前理解纳入我们的人类IRL模型,以便机器人可以选择最大化人类理解的演示。我们还提出了一种新颖的措施,以估算人类在看不见环境中预测机器人行为实例的困难。一项用户研究发现,我们的测试难度与人类绩效和信心息息相关。有趣的是,选择人类的信念和反事实时,选择示范会在易于测试中降低人类绩效,但在困难测试中提高了性能,从而提供了有关如何最好地利用此类模型的见解。
To collaborate well with robots, we must be able to understand their decision making. Humans naturally infer other agents' beliefs and desires by reasoning about their observable behavior in a way that resembles inverse reinforcement learning (IRL). Thus, robots can convey their beliefs and desires by providing demonstrations that are informative for a human learner's IRL. An informative demonstration is one that differs strongly from the learner's expectations of what the robot will do given their current understanding of the robot's decision making. However, standard IRL does not model the learner's existing expectations, and thus cannot do this counterfactual reasoning. We propose to incorporate the learner's current understanding of the robot's decision making into our model of human IRL, so that a robot can select demonstrations that maximize the human's understanding. We also propose a novel measure for estimating the difficulty for a human to predict instances of a robot's behavior in unseen environments. A user study finds that our test difficulty measure correlates well with human performance and confidence. Interestingly, considering human beliefs and counterfactuals when selecting demonstrations decreases human performance on easy tests, but increases performance on difficult tests, providing insight on how to best utilize such models.