基于多智体学习的多小区NOMA协作波束训练

Multi-cell NOMA cooperative beam training based on multi-agent learning

  • 摘要: 该文主要研究毫米波网络中协作非正交多址(non-orthogonal multiple access,NOMA)下的多小区的波束赋形优化问题。为了最大化系统吞吐量,并且考虑用户位置及信道信息,将基站的波束配置问题建模为马尔竞争博弈问题,并采用强化学习算法多智体深度确定性策略梯度(multi-agent deep deterministic policy gradient,MADDPG)对其求解,设计了一种多智能体强化学习的多小区协作NOMA波束赋形训练算法,以合理分配多基站体系中的波束、功率等资源,并提高系统的吞吐量。仿真结果表明,提出的MADDPG算法能达到更好的系统吞吐量及用户覆盖率。

     

    Abstract: This paper mainly focuses on the beamforming training problem in cooperative non-orthogonal multiple access (NOMA) scenarios in millimeter-wave communication, extending the work from single-cell NOMA to multi-cell NOMA scenarios. To maximize system throughput while considering user locations and channel information, the beam configuration problem at the base station is modeled as a Markov cooperative-competitive game problem. And then the problem is solved by exploiting multi-agent deep deterministic policy gradient (MADDPG) based reinforcement learning algorithm. A multi-agent reinforcement learning-based beamforming training algorithm for cooperative NOMA in multi-cell scenarios is designed to effectively allocate resources such as beams and power in multi-base station systems, thereby enhancing system throughput. Numerical simulations demonstrate that the proposed MADDPG algorithm achieves better system throughput and user coverage.

     

/

返回文章
返回
Baidu
map