TY - JOUR
T1 - Blockchain-Assisted Demonstration Cloning for Multiagent Deep Reinforcement Learning
AU - Alagha, Ahmed
AU - Bentahar, Jamal
AU - Otrok, Hadi
AU - Singh, Shakti
AU - Mizouni, Rabeb
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024/3/1
Y1 - 2024/3/1
N2 - Multiagent deep reinforcement learning (MDRL) is a promising research area in which agents learn complex behaviors in cooperative or competitive environments. However, MDRL comes with several challenges that hinder its usability, including sample efficiency, curse of dimensionality, and environment exploration. Recent works proposing federated reinforcement learning (FRL) to tackle these issues suffer from problems related to model restrictions and maliciousness. Other proposals using reward shaping (RS) require considerable engineering and could lead to local optima. In this article, we propose a novel Blockchain-assisted multiexpert demonstration cloning (MEDC) framework for MDRL. The proposed method utilizes expert demonstrations in guiding the learning of new MDRL agents, by suggesting exploration actions in the environment. A model sharing framework on Blockchain is designed to allow users to share their trained models, which can be allocated as expert models to requesting users to aid in training MDRL systems. A Consortium Blockchain is adopted to enable traceable and autonomous execution without the need for a single trusted entity. Smart Contracts are designed to manage users and models allocation, which are shared using IPFS. The proposed framework is tested on several applications and is benchmarked against existing methods in FRL, RS, and imitation learning-assisted RL. The results show the outperformance of the proposed framework in terms of learning speed and resiliency to faulty and malicious models.
AB - Multiagent deep reinforcement learning (MDRL) is a promising research area in which agents learn complex behaviors in cooperative or competitive environments. However, MDRL comes with several challenges that hinder its usability, including sample efficiency, curse of dimensionality, and environment exploration. Recent works proposing federated reinforcement learning (FRL) to tackle these issues suffer from problems related to model restrictions and maliciousness. Other proposals using reward shaping (RS) require considerable engineering and could lead to local optima. In this article, we propose a novel Blockchain-assisted multiexpert demonstration cloning (MEDC) framework for MDRL. The proposed method utilizes expert demonstrations in guiding the learning of new MDRL agents, by suggesting exploration actions in the environment. A model sharing framework on Blockchain is designed to allow users to share their trained models, which can be allocated as expert models to requesting users to aid in training MDRL systems. A Consortium Blockchain is adopted to enable traceable and autonomous execution without the need for a single trusted entity. Smart Contracts are designed to manage users and models allocation, which are shared using IPFS. The proposed framework is tested on several applications and is benchmarked against existing methods in FRL, RS, and imitation learning-assisted RL. The results show the outperformance of the proposed framework in terms of learning speed and resiliency to faulty and malicious models.
KW - Blockchain
KW - demonstration cloning
KW - imitation learning (IL)
KW - multiagent deep reinforcement learning (MDRL)
KW - proximal policy optimization (PPO)
KW - smart contracts
UR - http://www.scopus.com/inward/record.url?scp=85171576781&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2023.3316078
DO - 10.1109/JIOT.2023.3316078
M3 - Article
AN - SCOPUS:85171576781
SN - 2327-4662
VL - 11
SP - 7710
EP - 7723
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 5
ER -