TY - JOUR
T1 - Cooperative Multiagent Deep Reinforcement Learning Methods for UAV-Aided Mobile Edge Computing Networks
AU - Kim, Mintae
AU - Lee, Hoon
AU - Hwang, Sangwon
AU - Debbah, Merouane
AU - Lee, Inkyu
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024
Y1 - 2024
N2 - This article presents a cooperative multiagent deep reinforcement learning (MADRL) approach for unmanned aerial vehicle (UAV)-aided mobile edge computing (MEC) networks. An UAV with computing capability can provide task offlaoding services to ground Internet of Things devices (IDs). With partial observation of the entire network state, the UAV and the IDs individually determine their MEC strategies, i.e., UAV trajectory, resource allocation, and task offloading policy. This requires joint optimization of decision-making process and coordination strategies among the UAV and the IDs. To address this difficulty, the proposed cooperative MADRL approach computes two types of action variables, namely, message action and solution action, each of which is generated by dedicated actor neural networks (NNs). As a result, each agent can automatically encapsulate its coordination messages to enhance the MEC performance in the decentralized manner. The proposed actor structure is designed based on graph attention networks such that operations are possible regardless of the number of IDs. A scalable training algorithm is also proposed to train a group of NNs for arbitrary network configurations. Numerical results demonstrate the superiority of the proposed cooperative MADRL approach over conventional methods.
AB - This article presents a cooperative multiagent deep reinforcement learning (MADRL) approach for unmanned aerial vehicle (UAV)-aided mobile edge computing (MEC) networks. An UAV with computing capability can provide task offlaoding services to ground Internet of Things devices (IDs). With partial observation of the entire network state, the UAV and the IDs individually determine their MEC strategies, i.e., UAV trajectory, resource allocation, and task offloading policy. This requires joint optimization of decision-making process and coordination strategies among the UAV and the IDs. To address this difficulty, the proposed cooperative MADRL approach computes two types of action variables, namely, message action and solution action, each of which is generated by dedicated actor neural networks (NNs). As a result, each agent can automatically encapsulate its coordination messages to enhance the MEC performance in the decentralized manner. The proposed actor structure is designed based on graph attention networks such that operations are possible regardless of the number of IDs. A scalable training algorithm is also proposed to train a group of NNs for arbitrary network configurations. Numerical results demonstrate the superiority of the proposed cooperative MADRL approach over conventional methods.
KW - Graph attention network (GAT)
KW - reinforcement learning
KW - unmanned aerial vehicle (UAV) mobile edge computing (MEC)
UR - https://www.scopus.com/pages/publications/85201772219
U2 - 10.1109/JIOT.2024.3447090
DO - 10.1109/JIOT.2024.3447090
M3 - Article
AN - SCOPUS:85201772219
SN - 2327-4662
VL - 11
SP - 38040
EP - 38053
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 23
ER -