论文标题
三盖式:具有生成对抗网的多模式模仿学习框架
Triple-GAIL: A Multi-Modal Imitation Learning Framework with Generative Adversarial Nets
论文作者
论文摘要
生成的对抗性模仿学习(GAIL)通过利用生成的对抗网,尤其是在机器人学习领域,显示出令人鼓舞的结果。但是,对孤立的单模式演示的要求限制了对现实世界情景的可扩展性,例如自动驾驶汽车对适当理解人类驾驶员行为的需求。在本文中,我们提出了一个新型的多模式Gail框架,名为Triple-Gail,该框架能够通过引入辅助技能选择器来从专家演示中共同学习技能选择和模仿数据增强目的的经验。我们分别为发电机和选择器的融合提供了理论保证。对真正的驱动程序轨迹和实时策略游戏数据集的实验表明,三盖式可以更好地拟合靠近示威者的多模式行为,并且表现优于最先进的方法。
Generative adversarial imitation learning (GAIL) has shown promising results by taking advantage of generative adversarial nets, especially in the field of robot learning. However, the requirement of isolated single modal demonstrations limits the scalability of the approach to real world scenarios such as autonomous vehicles' demand for a proper understanding of human drivers' behavior. In this paper, we propose a novel multi-modal GAIL framework, named Triple-GAIL, that is able to learn skill selection and imitation jointly from both expert demonstrations and continuously generated experiences with data augmentation purpose by introducing an auxiliary skill selector. We provide theoretical guarantees on the convergence to optima for both of the generator and the selector respectively. Experiments on real driver trajectories and real-time strategy game datasets demonstrate that Triple-GAIL can better fit multi-modal behaviors close to the demonstrators and outperforms state-of-the-art methods.