TY - JOUR
T1 - Cooperative motion planning and control of a group of autonomous underwater vehicles using twin-delayed deep deterministic policy gradient
AU - Hadi, Behnaz
AU - Khosravi, Alireza
AU - Sarhadi, Pouria
N1 - © 2024 Elsevier Ltd. All rights reserved. This is the accepted manuscript version of an article which has been published in final form at https://doi.org/10.1016/j.apor.2024.103977
PY - 2024/6/30
Y1 - 2024/6/30
N2 - The cooperative execution of complex tasks can lead to desirable outcomes and increase the likelihood of mission success. Nevertheless, coordinating the movements of multiple autonomous underwater vehicles (AUVs) in a collaborative manner is challenging due to nonlinear dynamics and environmental disturbances. The paper presents a decentralized deep reinforcement learning algorithm for AUVs that enables cooperative motion planning and obstacle avoidance. The goal is to formulate control policies for AUVs, empowering each vehicle to create its optimal collision-free path through adjustments in speed and heading. To ensure safe navigation of multiple AUVs, COLlision AVoidance (COLAV) plays a crucial role. Therefore, the implementation of a multi-layer region control strategy enhances the AUVs’ responsiveness to nearby obstacles, leading to improved COLAV. Furthermore, a reward function is formulated to consider four criteria: path planning, obstacle- and self-COLAV, as well as feasible control signals, with the aim of strengthening the proposed strategy. Notably, the devised scheme demonstrates robustness against disturbances A comparative study is conducted with the well-established Artificial Potential Field (APF) planning method. The simulation results indicate that the proposed system effectively and safely guides the AUVs to their goals and exhibits desirable generalizability.
AB - The cooperative execution of complex tasks can lead to desirable outcomes and increase the likelihood of mission success. Nevertheless, coordinating the movements of multiple autonomous underwater vehicles (AUVs) in a collaborative manner is challenging due to nonlinear dynamics and environmental disturbances. The paper presents a decentralized deep reinforcement learning algorithm for AUVs that enables cooperative motion planning and obstacle avoidance. The goal is to formulate control policies for AUVs, empowering each vehicle to create its optimal collision-free path through adjustments in speed and heading. To ensure safe navigation of multiple AUVs, COLlision AVoidance (COLAV) plays a crucial role. Therefore, the implementation of a multi-layer region control strategy enhances the AUVs’ responsiveness to nearby obstacles, leading to improved COLAV. Furthermore, a reward function is formulated to consider four criteria: path planning, obstacle- and self-COLAV, as well as feasible control signals, with the aim of strengthening the proposed strategy. Notably, the devised scheme demonstrates robustness against disturbances A comparative study is conducted with the well-established Artificial Potential Field (APF) planning method. The simulation results indicate that the proposed system effectively and safely guides the AUVs to their goals and exhibits desirable generalizability.
KW - Multi-AUVs
KW - Motion planning
KW - Obstacle avoidance
KW - Ocean current
KW - Deep reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85189932571&partnerID=8YFLogxK
U2 - 10.1016/j.apor.2024.103977
DO - 10.1016/j.apor.2024.103977
M3 - Article
SN - 0141-1187
VL - 147
SP - 1
EP - 12
JO - Applied Ocean Research
JF - Applied Ocean Research
M1 - 103977
ER -