This study delves into the application of a deep Q-Network (DQN) algorithm,which integrates strategies of Navigational Priority (NP) and Prioritized Experience Replay (PER),for intelligent path planning in maritime environments. Unlike conventional path planning algorithms,our optimized model autonomously explores and learns the patterns of the maritime environment without relying on manually constructed global maritime information. We have developed a maritime simulation environment based on the Gym framework to simulate and validate our improved DQN model. This model incorporates the mechanisms of Navigational Priority and Prioritized Experience Replay,enhancing the algorithm's learning efficiency for critical decisions by adjusting the frequency of experience sample utilization during the learning process. Additionally,the introduction of a novel reward function has further strengthened the model's adaptability and stability in addressing path planning issues. Simulation experiments demonstrate that our model significantly outperforms baseline methods in avoiding obstacles and finding optimal routes,showcasing notable generalizability and exceptional stability.
关键词
改进深度Q网络/海上模拟仿真环境/航海优先级/奖赏函数
Key words
improved deep Q-Network/maritime simulation environment/navigational priority/reward function