论文标题
部分可观测时空混沌系统的无模型预测
Cognitive Models as Simulators: The Case of Moral Decision-Making
论文作者
论文摘要
为了实现理想的性能,当前的AI系统通常需要大量的培训数据。这在收集数据既昂贵又耗时的域中尤其有问题,例如,AI系统需要与人类进行多种互动,从中收集反馈。在这项工作中,我们证实了$ \ textIt {认知模型为模拟器} $的想法,这将使AI系统与认知模型而不是人类相互作用并收集反馈,从而使他们的培训过程较低,更快。在这里,我们通过与《终极游戏》(UG)的认知模型(UG)互动,在道德决策的背景下通过道德决策来利用这一想法,这是行为和脑科学领域的规范任务,以研究公平。有趣的是,这些RL代理学会根据模拟的UG响应者的情绪状态理性地适应其行为。我们的工作表明,使用认知模型作为人类的模拟者是训练AI系统的有效方法,为计算认知科学提供了对AI做出贡献的重要方法。
To achieve desirable performance, current AI systems often require huge amounts of training data. This is especially problematic in domains where collecting data is both expensive and time-consuming, e.g., where AI systems require having numerous interactions with humans, collecting feedback from them. In this work, we substantiate the idea of $\textit{cognitive models as simulators}$, which is to have AI systems interact with, and collect feedback from, cognitive models instead of humans, thereby making their training process both less costly and faster. Here, we leverage this idea in the context of moral decision-making, by having reinforcement learning (RL) agents learn about fairness through interacting with a cognitive model of the Ultimatum Game (UG), a canonical task in behavioral and brain sciences for studying fairness. Interestingly, these RL agents learn to rationally adapt their behavior depending on the emotional state of their simulated UG responder. Our work suggests that using cognitive models as simulators of humans is an effective approach for training AI systems, presenting an important way for computational cognitive science to make contributions to AI.