TY - JOUR
T1 - Resource Allocation for NOMA-MEC Systems in Ultra-Dense Networks
T2 - A Learning Aided Mean-Field Game Approach
AU - Li, Lixin
AU - Cheng, Qianqian
AU - Tang, Xiao
AU - Bai, Tong
AU - Chen, Wei
AU - Ding, Zhiguo
AU - Han, Zhu
N1 - Publisher Copyright:
© 2002-2012 IEEE.
PY - 2021/3
Y1 - 2021/3
N2 - Attracted by the advantages of multi-access edge computing (MEC) and non-orthogonal multiple access (NOMA), this article studies the resource allocation problem of a NOMA-MEC system in an ultra-dense network (UDN), where each user may opt for offloading tasks to the MEC server when it is computationally intensive. Our optimization goal is to minimize the system computation cost, concerning the energy consumption and task delay of users. In order to tackle the non-convexity issue of the objective function, we decouple this problem into two sub-problems: user clustering as well as jointly power and computation resource allocation. Firstly, we propose a user clustering matching (UCM) algorithm exploiting the differences in channel gains of users. Then, relying on the mean-field game (MFG) framework, we solve the resource allocation problem for intensive user deployment, using the novel deep deterministic policy gradient (DDPG) method, which is termed by a mean-field-deep deterministic policy gradient (MF-DDPG) algorithm. Finally, a jointly iterative optimization algorithm (JIOA) of UCM and MF-DDPG is proposed to minimize the computation cost of users. The simulation results demonstrate that the proposed algorithm exhibits rapid convergence, and is capable of efficiently reducing both the energy consumption and task delay of users.
AB - Attracted by the advantages of multi-access edge computing (MEC) and non-orthogonal multiple access (NOMA), this article studies the resource allocation problem of a NOMA-MEC system in an ultra-dense network (UDN), where each user may opt for offloading tasks to the MEC server when it is computationally intensive. Our optimization goal is to minimize the system computation cost, concerning the energy consumption and task delay of users. In order to tackle the non-convexity issue of the objective function, we decouple this problem into two sub-problems: user clustering as well as jointly power and computation resource allocation. Firstly, we propose a user clustering matching (UCM) algorithm exploiting the differences in channel gains of users. Then, relying on the mean-field game (MFG) framework, we solve the resource allocation problem for intensive user deployment, using the novel deep deterministic policy gradient (DDPG) method, which is termed by a mean-field-deep deterministic policy gradient (MF-DDPG) algorithm. Finally, a jointly iterative optimization algorithm (JIOA) of UCM and MF-DDPG is proposed to minimize the computation cost of users. The simulation results demonstrate that the proposed algorithm exhibits rapid convergence, and is capable of efficiently reducing both the energy consumption and task delay of users.
KW - deep deterministic policy gradient
KW - deep reinforcement learning
KW - mean-field game
KW - Multi-access edge computing
KW - non-orthogonal multiple access
UR - http://www.scopus.com/inward/record.url?scp=85102726067&partnerID=8YFLogxK
U2 - 10.1109/TWC.2020.3033843
DO - 10.1109/TWC.2020.3033843
M3 - Article
AN - SCOPUS:85102726067
SN - 1536-1276
VL - 20
SP - 1487
EP - 1500
JO - IEEE Transactions on Wireless Communications
JF - IEEE Transactions on Wireless Communications
IS - 3
M1 - 9247446
ER -