论文标题
任务诱导的表示学习
Task-Induced Representation Learning
论文作者
论文摘要
在这项工作中,我们评估了在视觉上复杂环境中对决策做出的代表性学习方法的有效性。表示学习对于从高维输入中的有效加强学习(RL)至关重要。基于重建,预测或对比度学习的无监督表示学习方法已显示出很大的学习效率提高。但是,它们主要在清洁实验室或模拟环境中进行了评估。相比之下,实际环境在视觉上是复杂的,并且包含大量混乱和干扰物。无监督的表示形式将学会为这些干扰者建模,并可能损害代理商的学习效率。相比之下,我们称任务诱导的表示形式学习的替代方法,利用任务信息(例如奖励或先前任务的演示),以关注场景中与任务相关的部分并忽略干扰器。我们研究了四种视觉上复杂环境的无监督和任务诱导的表示方法的有效性,从分散DMCONTROL到CARLA驾驶模拟器。对于RL和模仿学习,我们都会发现,即使在视觉上复杂的场景中,代表性学习也通常可以提高看不见的任务的样本效率,而任务引起的表示形式与无监督的替代方案相比可以双倍地学习效率。代码可在https://clvrai.com/tarp上找到。
In this work, we evaluate the effectiveness of representation learning approaches for decision making in visually complex environments. Representation learning is essential for effective reinforcement learning (RL) from high-dimensional inputs. Unsupervised representation learning approaches based on reconstruction, prediction or contrastive learning have shown substantial learning efficiency gains. Yet, they have mostly been evaluated in clean laboratory or simulated settings. In contrast, real environments are visually complex and contain substantial amounts of clutter and distractors. Unsupervised representations will learn to model such distractors, potentially impairing the agent's learning efficiency. In contrast, an alternative class of approaches, which we call task-induced representation learning, leverages task information such as rewards or demonstrations from prior tasks to focus on task-relevant parts of the scene and ignore distractors. We investigate the effectiveness of unsupervised and task-induced representation learning approaches on four visually complex environments, from Distracting DMControl to the CARLA driving simulator. For both, RL and imitation learning, we find that representation learning generally improves sample efficiency on unseen tasks even in visually complex scenes and that task-induced representations can double learning efficiency compared to unsupervised alternatives. Code is available at https://clvrai.com/tarp.