Journal of Marine Science and Engineering | |
AUV Obstacle Avoidance Planning Based on Deep Reinforcement Learning | |
Hongjian Wang1  Dan Yu1  Chengfeng Li1  Jianya Yuan1  Honghan Zhang1  Changjian Lin2  | |
[1] College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 045100, China;School of Information and Control Engineering, China University of Mining and Technology, Xuzhou 221000, China; | |
关键词: autonomous underwater vehicle (AUV); collision avoidance planning; deep reinforcement learning (DRL); double-DQN (D-DQN); | |
DOI : 10.3390/jmse9111166 | |
来源: DOAJ |
【 摘 要 】
In a complex underwater environment, finding a viable, collision-free path for an autonomous underwater vehicle (AUV) is a challenging task. The purpose of this paper is to establish a safe, real-time, and robust method of collision avoidance that improves the autonomy of AUVs. We propose a method based on active sonar, which utilizes a deep reinforcement learning algorithm to learn the processed sonar information to navigate the AUV in an uncertain environment. We compare the performance of double deep Q-network algorithms with that of a genetic algorithm and deep learning. We propose a line-of-sight guidance method to mitigate abrupt changes in the yaw direction and smooth the heading changes when the AUV switches trajectory. The different experimental results show that the double deep Q-network algorithms ensure excellent collision avoidance performance. The effectiveness of the algorithm proposed in this paper was verified in three environments: random static, mixed static, and complex dynamic. The results show that the proposed algorithm has significant advantages over other algorithms in terms of success rate, collision avoidance performance, and generalization ability. The double deep Q-network algorithm proposed in this paper is superior to the genetic algorithm and deep learning in terms of the running time, total path, performance in avoiding collisions with moving obstacles, and planning time for each step. After the algorithm is trained in a simulated environment, it can still perform online learning according to the information of the environment after deployment and adjust the weight of the network in real-time. These results demonstrate that the proposed approach has significant potential for practical applications.
【 授权许可】
Unknown