Journal of System Simulation ›› 2025, Vol. 37 ›› Issue (11): 2754-2767.doi: 10.16182/j.issn1004731x.joss.24-0678

• Papers • Previous Articles    

AUV Path Planning Based on Behavior Cloning and Improved DQN in Partially Unknown Environments

Xing Lijing1, Li Min1, Zeng Xiangguang1, Zhang Ping2, Peng Bei2   

  1. 1.School of Mechanical Engineering, Southwest Jiaotong University, Chengdu 610031, China
    2.University of Electronic Science and Technology of China, Chengdu 610031, China
  • Received:2024-06-26 Revised:2024-09-11 Online:2025-11-18 Published:2025-11-27
  • Contact: Li Min

Abstract:

To address the problems of large randomness and slow convergence of the DQN dynamic path planning algorithm for a single autonomous underwater vehicle (AUV) in a partially unknown environment, a path planning method combining behavior cloning with A* algorithm and DQN (BA_DQN) was proposed. Based on the known environmental information, an improved A* algorithm incorporating ocean current resistance was proposed to guide DQN, thereby reducing the randomness of the DQN algorithm. By considering the complexity of the marine environment, the sampling probability was improved again after expanding the positive experience pool to enhance the training success rate. To address the problem of slow convergence in DQN, an improved algorithm based on reinforcement learning followed by behavior cloning was proposed. The BA_DQN was used to control AUV pathfinding, and simulation experiments were carried out in different task scenarios. The simulation results show that the training time of the BA_DQN algorithm is shorter than that of the DQN algorithm; its decision-making is faster than that of the A* algorithm, and its sailing time is shorter.

Key words: AUV, path planning, A* algorithm, reinforcement learning, behavior cloning

CLC Number: