首页|Multi-Agent Path Planning Method Based on Improved Deep Q-Network in Dynamic Environments

Multi-Agent Path Planning Method Based on Improved Deep Q-Network in Dynamic Environments

扫码查看
The multi-agent path planning problem presents significant challenges in dynamic environments,pri-marily due to the ever-changing positions of obstacles and the complex interactions between agents'actions.These factors contribute to a tendency for the solution to converge slowly,and in some cases,diverge altogether.In addressing this issue,this paper introduces a novel approach utilizing a double dueling deep Q-network(D3QN),tailored for dynamic multi-agent environments.A novel reward function based on multi-agent positional con-straints is designed,and a training strategy based on incremental learning is performed to achieve collaborative path planning of multiple agents.Moreover,the greedy and Boltzmann probability selection policy is introduced for action selection and avoiding convergence to local extremum.To match radar and image sensors,a convolu-tional neural network-long short-term memory(CNN-LSTM)architecture is constructed to extract the feature of multi-source measurement as the input of the D3QN.The algorithm's efficacy and reliability are validated in a simulated environment,utilizing robot operating system and Gazebo.The simulation results show that the proposed algorithm provides a real-time solution for path planning tasks in dynamic scenarios.In terms of the average success rate and accuracy,the proposed method is superior to other deep learning algorithms,and the convergence speed is also improved.

multi-agentpath planningdeep reinforcement learningdeep Q-network

李舒逸、李旻哲、敬忠良

展开 >

School of Aeronautics and Astronautics,Shanghai Jiao Tong University,Shanghai 200240,China

2024

上海交通大学学报(英文版)
上海交通大学

上海交通大学学报(英文版)

影响因子:0.151
ISSN:1007-1172
年,卷(期):2024.29(4)