首页|融合人工势场的Q_learning算法移动机器人路径规划

融合人工势场的Q_learning算法移动机器人路径规划

扫码查看
传统的Q_learning算法在路径规划中存在收敛速度慢和容易陷入局部最优解等问题,为此,提出一种融合人工势场法和Q_learning算法的路径规划算法。通过引入势场值来优化选择状态时的奖励值,在栅格地图中构建虚拟势场,并在初始化Q值函数时引入人工势场的引力函数,在更新Q值函数时引入斥力函数。最后,通过仿真模拟实验对比了传统算法和改进后算法的性能。实验结果表明:改进后的算法收敛时间减少了 21。5%,规划的路径更平滑,验证了该方法的有效性。
Mobile Robot Path Planning with Integrated Artificial Potential Fields and Q_learning Algorithm
A path planning algorithm that integrates artificial potential field methods with the traditional Q_learning algorithm was proposed to address the issues of slow convergence and the propensity to fall into local optima.The potential field values were introduced to optimize the reward values during state selection,and a virtual potential field was constructed within a grid map.The gravitational function of the artificial potential field was incorporated when initializing the Q-value function,while the repulsive function was intro-duced during the updating of the Q-value function.Finally,the performance of the traditional algorithm and the improved algorithm was compared by simulation experiments.The experimental results show that the convergence time of the improved algorithm is reduced by 21.5%,the planned paths are smoother,and the effectiveness of the proposed method is verified.

path planningartificial potential field methodQ_learningreward values

张雨晴、高金凤、苏雯、潘海鹏

展开 >

浙江理工大学信息科学与工程学院,浙江 杭州 310018

路径规划 人工势场法 Q_learning 奖励值

2024

机床与液压
中国机械工程学会 广州机械科学研究院有限公司

机床与液压

CSTPCD北大核心
影响因子:0.32
ISSN:1001-3881
年,卷(期):2024.52(23)