论文标题
未知环境映射的异构多机构增强学习
Heterogeneous Multi-Agent Reinforcement Learning for Unknown Environment Mapping
论文作者
论文摘要
异质多代理场景中的强化学习对于现实世界的应用很重要,但提出了挑战,而不是在均质环境和简单基准中看到的挑战。在这项工作中,我们提出了一种参与者批评算法,该算法允许一组异质的代理人学习分散的控制政策,以覆盖未知环境。这项任务是国家安全和应急组织感兴趣的,希望通过部署无人驾驶飞机团队来提高危险地区的情境意识。为了在未知环境中解决这个多代理的覆盖路径计划问题,我们扩大了具有新的状态编码结构和三胞胎学习损失的多代理参与者批判性架构,以支持异质剂学习。我们开发了一个模拟环境,其中包括现实世界中的环境因素,例如湍流,沟通延迟和代理损失,以训练代理团队,并探究它们的鲁棒性和灵活性,以实现这种干扰。
Reinforcement learning in heterogeneous multi-agent scenarios is important for real-world applications but presents challenges beyond those seen in homogeneous settings and simple benchmarks. In this work, we present an actor-critic algorithm that allows a team of heterogeneous agents to learn decentralized control policies for covering an unknown environment. This task is of interest to national security and emergency response organizations that would like to enhance situational awareness in hazardous areas by deploying teams of unmanned aerial vehicles. To solve this multi-agent coverage path planning problem in unknown environments, we augment a multi-agent actor-critic architecture with a new state encoding structure and triplet learning loss to support heterogeneous agent learning. We developed a simulation environment that includes real-world environmental factors such as turbulence, delayed communication, and agent loss, to train teams of agents as well as probe their robustness and flexibility to such disturbances.