TY - JOUR
T1 - Enhancing collaboration in uncertain environment
T2 - Multi-Agent Reinforcement Learning for underwater monitoring
AU - Luvisutto, Alberto
AU - Celani, Antonio
AU - Renda, Federico
AU - Stefanini, Cesare
AU - De Masi, Giulia
N1 - Publisher Copyright:
© 2025 The Author(s)
PY - 2025/6/5
Y1 - 2025/6/5
N2 - Underwater monitoring is extremely complex due to the lack of a global localization system, limited communication and environmental factors such as turbidity and darkness that limit visibility, affecting control and situational awareness. Typically, monitoring relies on a single autonomous underwater vehicle (AUV) or a set of independent AUVs; techniques which are prone to failure as they rely only on onboard odometry and sensors, making missions vulnerable to malfunctions, damage, and noise. To address these challenges, we propose a Multi-Agent Reinforcement Learning (MARL) framework to enable cooperation among multiple AUVs, mitigating the limitations of the underwater environment. Our in-silico solution focuses on a group of robots learning a strategy to follow a partially hidden underwater pipe without global localization, while dealing with environmental disturbances affecting sensors and actuators. The numerosity of the agents, and most importantly their collaboration, helps overcome underwater visibility constraints. By sharing relative position information of neighboring agents with respect to the pipe, navigation is improved. By introducing quantitative measures for pipe exploration, we show that cooperation significantly enhances system performance compared to independent agents. Emerging collaboration among robots allows the swarm to complete pipe inspections faster and more efficiently than non-cooperative baseline models of non-interacting agents, even under extremely reduced visibility scenarios. Moreover, single agents also benefit from cooperation, learning effective policies more quickly and covering a longer portion of the pipe. Finally, our model guarantees explainability. We analyze learned strategies and provide a visualization method that allows the interpretation of the learned policies.
AB - Underwater monitoring is extremely complex due to the lack of a global localization system, limited communication and environmental factors such as turbidity and darkness that limit visibility, affecting control and situational awareness. Typically, monitoring relies on a single autonomous underwater vehicle (AUV) or a set of independent AUVs; techniques which are prone to failure as they rely only on onboard odometry and sensors, making missions vulnerable to malfunctions, damage, and noise. To address these challenges, we propose a Multi-Agent Reinforcement Learning (MARL) framework to enable cooperation among multiple AUVs, mitigating the limitations of the underwater environment. Our in-silico solution focuses on a group of robots learning a strategy to follow a partially hidden underwater pipe without global localization, while dealing with environmental disturbances affecting sensors and actuators. The numerosity of the agents, and most importantly their collaboration, helps overcome underwater visibility constraints. By sharing relative position information of neighboring agents with respect to the pipe, navigation is improved. By introducing quantitative measures for pipe exploration, we show that cooperation significantly enhances system performance compared to independent agents. Emerging collaboration among robots allows the swarm to complete pipe inspections faster and more efficiently than non-cooperative baseline models of non-interacting agents, even under extremely reduced visibility scenarios. Moreover, single agents also benefit from cooperation, learning effective policies more quickly and covering a longer portion of the pipe. Finally, our model guarantees explainability. We analyze learned strategies and provide a visualization method that allows the interpretation of the learned policies.
KW - MARL
KW - Multi-agent
KW - Reinforcement Learning
KW - Underwater swarm
UR - https://www.scopus.com/pages/publications/105001040660
U2 - 10.1016/j.eswa.2025.127256
DO - 10.1016/j.eswa.2025.127256
M3 - Article
AN - SCOPUS:105001040660
SN - 0957-4174
VL - 277
JO - Expert Systems with Applications
JF - Expert Systems with Applications
M1 - 127256
ER -