论文标题

团体代理增强学习

Group-Agent Reinforcement Learning

论文作者

Wu, Kaiyue, Zeng, Xiao-Jun

论文摘要

如果多个地理分布式代理合作执行其单独的RL任务,则它可以在很大程度上受益于每个代理的增强学习(RL)过程。不同于多代理的增强学习(MARL),其中多个代理在共同的环境中,应该学会合作或互相竞争,在这种情况下,在这种情况下,每个代理人都有其独立的环境,并且只能与他人进行交流以共享知识,而无需任何合作或竞争行为作为学习成果。实际上,这种情况在现实生活中广泛存在,它们的概念可以在许多应用中使用,但尚未得到充分理解,并且尚未得到很好的理解。作为第一项努力,我们提出了RL的群体代理系统,作为该方案的公式以及第三类RL系统,相对于单代理和多代理系统。然后,我们提出了一个称为DDAL(分散分布的异步学习)的分布式RL框架,该框架专为团体代理增强学习(GARL)而设计。我们通过实验表明,DDAL通过非常稳定的训练实现了理想的性能,并且具有良好的可扩展性。

It can largely benefit the reinforcement learning (RL) process of each agent if multiple geographically distributed agents perform their separate RL tasks cooperatively. Different from multi-agent reinforcement learning (MARL) where multiple agents are in a common environment and should learn to cooperate or compete with each other, in this case each agent has its separate environment and only communicates with others to share knowledge without any cooperative or competitive behaviour as a learning outcome. In fact, this scenario exists widely in real life whose concept can be utilised in many applications, but is not well understood yet and not well formulated. As the first effort, we propose group-agent system for RL as a formulation of this scenario and the third type of RL system with respect to single-agent and multi-agent systems. We then propose a distributed RL framework called DDAL (Decentralised Distributed Asynchronous Learning) designed for group-agent reinforcement learning (GARL). We show through experiments that DDAL achieved desirable performance with very stable training and has good scalability.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源