TY - JOUR
T1 - Deep reinforcement learning for the computation offloading in MIMO-based Edge Computing
AU - Sadiki, Abdeladim
AU - Bentahar, Jamal
AU - Dssouli, Rachida
AU - En-Nouaary, Abdeslam
AU - Otrok, Hadi
N1 - Funding Information:
Jamal Bentahar and Rachida Dssouli are supported by the Natural Sciences and Engineering Research Council of Canada (NSERC) . Jamal Bentahar is also supported by the Department of national Defence of Canada (DnD, the IDEaS program) and by Mitacs .
Publisher Copyright:
© 2023 Elsevier B.V.
PY - 2023/3/15
Y1 - 2023/3/15
N2 - Multi-access Edge Computing (MEC) has recently emerged as a potential technology to serve the needs of mobile devices (MDs) in 5G and 6G cellular networks. By offloading tasks to high-performance servers installed at the edge of the wireless networks, resource-limited MDs can cope with the proliferation of the recent computationally-intensive applications. In this paper, we study the computation offloading problem in a massive multiple-input multiple-output (MIMO)-based MEC system where the base stations are equipped with a large number of antennas. Our objective is to minimize the power consumption and offloading delay at the MDs under the stochastic system environment. To this end, we introduce new formulation of the problem as a Markov Decision Process (MDP) and propose two Deep Reinforcement Learning (DRL) algorithms to learn the optimal offloading policy without any prior knowledge of the environment dynamics. First, a Deep Q-Network (DQN)-based algorithm to solve the curse of the state space explosion is defined. Then, a more general Proximal Policy Optimization (PPO)-based algorithm to solve the problem of discrete action space is introduced. Simulation results show that our DRL-based solutions outperform the state-of-the-art algorithms. Moreover, our PPO algorithm exhibits stable performance and efficient offloading results compared to the benchmarks DQN and Double DQN (DDQN) strategies.
AB - Multi-access Edge Computing (MEC) has recently emerged as a potential technology to serve the needs of mobile devices (MDs) in 5G and 6G cellular networks. By offloading tasks to high-performance servers installed at the edge of the wireless networks, resource-limited MDs can cope with the proliferation of the recent computationally-intensive applications. In this paper, we study the computation offloading problem in a massive multiple-input multiple-output (MIMO)-based MEC system where the base stations are equipped with a large number of antennas. Our objective is to minimize the power consumption and offloading delay at the MDs under the stochastic system environment. To this end, we introduce new formulation of the problem as a Markov Decision Process (MDP) and propose two Deep Reinforcement Learning (DRL) algorithms to learn the optimal offloading policy without any prior knowledge of the environment dynamics. First, a Deep Q-Network (DQN)-based algorithm to solve the curse of the state space explosion is defined. Then, a more general Proximal Policy Optimization (PPO)-based algorithm to solve the problem of discrete action space is introduced. Simulation results show that our DRL-based solutions outperform the state-of-the-art algorithms. Moreover, our PPO algorithm exhibits stable performance and efficient offloading results compared to the benchmarks DQN and Double DQN (DDQN) strategies.
KW - Computation offloading
KW - Deep reinforcement learning
KW - Massive multiple-input multiple-output
KW - Multi-access Edge Computing
UR - http://www.scopus.com/inward/record.url?scp=85146627447&partnerID=8YFLogxK
U2 - 10.1016/j.adhoc.2022.103080
DO - 10.1016/j.adhoc.2022.103080
M3 - Article
AN - SCOPUS:85146627447
SN - 1570-8705
VL - 141
JO - Ad Hoc Networks
JF - Ad Hoc Networks
M1 - 103080
ER -