论文标题

蜂窝网络中关节频谱和功率分配的深度加强学习

Deep Reinforcement Learning for Joint Spectrum and Power Allocation in Cellular Networks

论文作者

Nasir, Yasar Sinan, Guo, Dongning

论文摘要

无线网络操作员通常将其拥有的无线电频谱分为多个子带。在蜂窝网络中,这些子带在许多细胞中都重复使用。为了减轻共通的干扰,通常制定了关节频谱和功率分配问题,以最大程度地提高总和率目标。解决此类问题的最著名算法通常需要瞬时全球通道状态信息和集中式优化器。实际上,这些算法尚未在具有随时间变化的子带的大型网络中实施。深度强化学习算法是解决复杂资源管理问题的有前途的工具。这里的一个主要挑战是频谱分配涉及离散的子带选择,而功率分配涉及连续变量。在本文中,提出了一个学习框架,以优化离散和连续决策变量。具体而言,两种独立的深钢筋学习算法被设计为同时执行和培训以最大化关节目标。仿真结果表明,所提出的方案的表现优于最先进的分数编程算法和基于深入强化学习的先前解决方案。

A wireless network operator typically divides the radio spectrum it possesses into a number of subbands. In a cellular network those subbands are then reused in many cells. To mitigate co-channel interference, a joint spectrum and power allocation problem is often formulated to maximize a sum-rate objective. The best known algorithms for solving such problems generally require instantaneous global channel state information and a centralized optimizer. In fact those algorithms have not been implemented in practice in large networks with time-varying subbands. Deep reinforcement learning algorithms are promising tools for solving complex resource management problems. A major challenge here is that spectrum allocation involves discrete subband selection, whereas power allocation involves continuous variables. In this paper, a learning framework is proposed to optimize both discrete and continuous decision variables. Specifically, two separate deep reinforcement learning algorithms are designed to be executed and trained simultaneously to maximize a joint objective. Simulation results show that the proposed scheme outperforms both the state-of-the-art fractional programming algorithm and a previous solution based on deep reinforcement learning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源