论文标题
TinyQmix:通过多代理增强学习的MMTC分布式访问控制
TinyQMIX: Distributed Access Control for mMTC via Multi-agent Reinforcement Learning
论文作者
论文摘要
分布式访问控制是大型机器类型通信(MMTC)的关键组件。在这种通信方案中,集中资源分配是不可扩展的,因为资源配置必须经常从基站发送到大量设备。我们在不依赖集中控制的情况下研究了用于选择资源的分布式增强学习。 MMTC的另一个重要特征是流量的零星和动态变化。关于分布式访问控制的现有研究假定流量负载是静态的,或者它们能够逐渐适应动态流量。我们通过培训TinyQmix(是一种轻巧的多代理深钢筋学习模型)来最大程度地减少适应期,以在部署前在各种流量模式下学习分布式的无线资源选择策略。因此,训练有素的代理能够快速适应动态流量并提供低访问延迟。提出数值结果以支持我们的主张。
Distributed access control is a crucial component for massive machine type communication (mMTC). In this communication scenario, centralized resource allocation is not scalable because resource configurations have to be sent frequently from the base station to a massive number of devices. We investigate distributed reinforcement learning for resource selection without relying on centralized control. Another important feature of mMTC is the sporadic and dynamic change of traffic. Existing studies on distributed access control assume that traffic load is static or they are able to gradually adapt to the dynamic traffic. We minimize the adaptation period by training TinyQMIX, which is a lightweight multi-agent deep reinforcement learning model, to learn a distributed wireless resource selection policy under various traffic patterns before deployment. Therefore, the trained agents are able to quickly adapt to dynamic traffic and provide low access delay. Numerical results are presented to support our claims.